• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 10:53
CEST 16:53
KST 23:53
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
[ASL19] Finals Recap: Standing Tall10HomeStory Cup 27 - Info & Preview18Classic wins Code S Season 2 (2025)16Code S RO4 & Finals Preview: herO, Rogue, Classic, GuMiho0TL Team Map Contest #5: Presented by Monster Energy6
Community News
$25,000 Streamerzone StarCraft Pro Series announced0Weekly Cups (June 30 - July 6): Classic Doubles5[BSL20] Non-Korean Championship 4x BSL + 4x China9Flash Announces Hiatus From ASL66Weekly Cups (June 23-29): Reynor in world title form?14
StarCraft 2
General
The GOAT ranking of GOAT rankings Weekly Cups (June 30 - July 6): Classic Doubles Jim claims he and Firefly were involved in match-fixing Esports World Cup 2025 - Final Player Roster Statistics for vetoed/disliked maps
Tourneys
$25,000 Streamerzone StarCraft Pro Series announced FEL Cracov 2025 (July 27) - $8000 live event Sparkling Tuna Cup - Weekly Open Tournament RSL: Revival, a new crowdfunded tournament series WardiTV Mondays
Strategy
How did i lose this ZvP, whats the proper response Simple Questions Simple Answers
Custom Maps
[UMS] Zillion Zerglings
External Content
Mutation # 481 Fear and Lava Mutation # 480 Moths to the Flame Mutation # 479 Worn Out Welcome Mutation # 478 Instant Karma
Brood War
General
ASL20 Preliminary Maps i aint gon lie to u bruh... BGH Auto Balance -> http://bghmmr.eu/ [ASL19] Finals Recap: Standing Tall SC uni coach streams logging into betting site
Tourneys
[BSL20] Non-Korean Championship 4x BSL + 4x China [BSL20] Grand Finals - Sunday 20:00 CET CSL Xiamen International Invitational The Casual Games of the Week Thread
Strategy
Simple Questions, Simple Answers I am doing this better than progamers do.
Other Games
General Games
Nintendo Switch Thread Stormgate/Frost Giant Megathread Path of Exile What do you want from future RTS games? Beyond All Reason
Dota 2
Official 'what is Dota anymore' discussion
League of Legends
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Heroes of StarCraft mini-set
TL Mafia
TL Mafia Community Thread Vanilla Mini Mafia
Community
General
US Politics Mega-thread Russo-Ukrainian War Thread Stop Killing Games - European Citizens Initiative Summer Games Done Quick 2025! Summer Games Done Quick 2024!
Fan Clubs
SKT1 Classic Fan Club! Maru Fan Club
Media & Entertainment
Anime Discussion Thread [Manga] One Piece [\m/] Heavy Metal Thread
Sports
2024 - 2025 Football Thread Formula 1 Discussion NBA General Discussion TeamLiquid Health and Fitness Initiative For 2023 NHL Playoffs 2024
World Cup 2022
Tech Support
Computer Build, Upgrade & Buying Resource Thread
TL Community
The Automated Ban List
Blogs
momentary artworks from des…
tankgirl
Culture Clash in Video Games…
TrAiDoS
from making sc maps to makin…
Husyelt
StarCraft improvement
iopq
Customize Sidebar...

Website Feedback

Closed Threads



Active: 718 users

AlphaStar AI goes 10-1 against human pros in demonstration…

Forum Index > SC2 General
374 CommentsPost a Reply
Prev 1 7 8 9 10 11 19 Next All
imp42
Profile Blog Joined November 2010
398 Posts
January 25 2019 15:47 GMT
#161
To sum up most of the debate so far:

- was the setup "fair" or not?
- did the AI play well or not?

In the chess community we were blown away by the games AlphaZero played about a year ago. We had never seen anything like it. However, it was also very disappointing to realize Deepmind wasn't interested in chess at all. It was nothing but a playing field to demonstrate the capabilities of their neural net. As soon as the experiment concluded successfully, the Deepmind team moved on and left the chess world wondering what could have been, if they had access. Imagine somebody allowing you to peek into a treasure chest full of amazing content, but then closes it and stores it away, not to be opened again.

Realistically we are in the same situation with StarCraft of course. Once Deepmind "beats the game" they will move on without missing a beat. But

"GOOD NEWS EVERYONE"
(Prof. Farnsworth)

Independent of the two questions above (fair setup? good play?) I think it's safe to say StarCraft still holds plenty of challenges for the AI. And even if at some point the peak of strategic depth has been reached in all matchups on arbitrary maps with a configurable average / max APM parameter, there is still the world of custom games to explore.

exciting!
50 pts Copper League
counting
Profile Joined January 2019
11 Posts
January 25 2019 15:51 GMT
#162
On January 26 2019 00:20 imp42 wrote:
Show nested quote +
On January 25 2019 21:14 Grumbels wrote:
On January 25 2019 13:00 counting wrote:
On January 25 2019 12:45 imp42 wrote:
On January 25 2019 12:14 vesicular wrote:
On January 25 2019 09:40 TheDougler wrote:
You don't know that it was the the camera change that actually was the determining factor here. It could be that Mana had a better idea of what he was up against. It could be that the warp prism threw off the AI's gameplan (I think it's this one). It could be that this AI isn't quite as good as other AIs.

[...]
The final I would say is to play only one agent. Every game used a different agent. It's akin to playing different players. TLO didn't know this when he was playing and played his matches as if it was the same agent and thus tried strats to counter what he just saw in the previous game, which of course didn't work. Playing against a single agent would be quite interesting.

A misconception IMO. There is no conceptual difference between "one agent" and "multiple agents", because you can simply combine x agents into one composite agent (which is exactly what they did).

Compare it to Innovation switching up his macro game with a 3-rax proxy cheese. It's not akin to playing different players, but the same player choosing a different game plan before he starts the game.

The concept of a composite agent gets interesting when you add a super-agent to it, responsible for picking a sub-agent to play a specific game in a boX match. I would imagine the super-agent would then be trained similar to a Texas Hold'em agent and converge to game-theoretical optima for cheese / standard ratio etc.


This actually has a technical term in machine learning community called ensemble learning. But I don't think it is that easy to implement as of yet. And for efficiency sake the single agent is actually very different from a group of agents which will absolutely require quite a bit of parallel processing to achieve (it is not as simple as installing more GPU can solve). And indeed these agents choose to represent the group of all agents in the AlphaStar league will be those encounter many different strategies and still win for the most part overall. It actually is a very difficult problem to introduce "novelty" and still able to adapt mid-game. The current system is simply not having any learning capability on the fly (within one game, in machine learning term, it is a system with offline learning, instead of active/online learning which is much much more difficult).

I don't know anything about AI, but wouldn't it be sufficient to simply have the bots play Bo5's against each other instead of Bo1's during the training phase? Because then they can still learn from what their opponent has been doing in previous games.

Well, you're not wrong. It's just that if you do that and actually want the bot to learn adaption patterns over multiple games, then you need to feed it the previously played games as input.

If you design that mechanism manually, the most simple approach I can think of is to feed it the history of wins/losses together with the respective agent as additional input:

Game 1: Agent "Mass Blinker" - loss
Game 2: Agent "Proxy Gates" - win
...

and so on (the agent names are chosen for illustration only - to the AI it would just be agent 1, agent 2, ...).

But if you don't want to do design anything manually = let the bot self-learn, then you'd have to feed the complete history of entire games as input, which blows up the input quite a bit.


There is already a matchmaking probabilities parameters in the reinforcement learning process as shown in the blog post
[image loading]

In evolutionary algorithm term, it is similar to mating or ranking mechanism put together, if an agent already played against certain agent with certain strategies, it shouldn't "mate" with the same agent as often, but really need to "mate/match" with newly introduced agents/variations, hence novelty search for new blood. However, the probability for playing against losing agents should still be higher I assume, so the policy learning will be able to reward a bit more and certain agents can solidify its sucessful strategies (try a few more times just to make sure so to speak)
BronzeKnee
Profile Joined March 2011
United States5217 Posts
Last Edited: 2019-01-25 16:11:31
January 25 2019 16:05 GMT
#163
I had been a proponent of oversaturating minerals for a long time after I saw Hister do it long ago. I gave up on it later, but I think I will go back to it.

There is so much to learn from an AI that doesn't care about norms or customs and does whatever it thinks gives it the best chance to win. As much as I try to avoid the pressure of following whatever everyone else does, its get me.

Deepmind has a very long way to go to beat Starcraft. A human can regularly take down the cheating SC2 AI with ease, I think Deepmind loses to the cheating AI at this point. It's micro isn't good enough to overcome the massive economic advantage the cheating AI gets.

And why didn't anyone cannon rush the AI? AI is always going to be weak to cheese. It will never play mind games better than a mind.
imp42
Profile Blog Joined November 2010
398 Posts
January 25 2019 16:20 GMT
#164
On January 26 2019 01:05 BronzeKnee wrote:
[...]
And why didn't anyone cannon rush the AI? AI is always going to be weak to cheese. It will never play mind games better than a mind.

My honest opinion? Because TLO and Mana were pulled aside in a quiet moment and politely asked to please tend towards macro games. We spotted some weakness in the defense, so trying to exploit that via a cannon rush would be a pretty straight-forward move...

just speculating though...
50 pts Copper League
BronzeKnee
Profile Joined March 2011
United States5217 Posts
Last Edited: 2019-01-25 16:22:51
January 25 2019 16:22 GMT
#165
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.
imp42
Profile Blog Joined November 2010
398 Posts
January 25 2019 16:24 GMT
#166
On January 26 2019 00:51 counting wrote:
Show nested quote +
On January 26 2019 00:20 imp42 wrote:
On January 25 2019 21:14 Grumbels wrote:
On January 25 2019 13:00 counting wrote:
On January 25 2019 12:45 imp42 wrote:
On January 25 2019 12:14 vesicular wrote:
On January 25 2019 09:40 TheDougler wrote:
You don't know that it was the the camera change that actually was the determining factor here. It could be that Mana had a better idea of what he was up against. It could be that the warp prism threw off the AI's gameplan (I think it's this one). It could be that this AI isn't quite as good as other AIs.

[...]
The final I would say is to play only one agent. Every game used a different agent. It's akin to playing different players. TLO didn't know this when he was playing and played his matches as if it was the same agent and thus tried strats to counter what he just saw in the previous game, which of course didn't work. Playing against a single agent would be quite interesting.

A misconception IMO. There is no conceptual difference between "one agent" and "multiple agents", because you can simply combine x agents into one composite agent (which is exactly what they did).

Compare it to Innovation switching up his macro game with a 3-rax proxy cheese. It's not akin to playing different players, but the same player choosing a different game plan before he starts the game.

The concept of a composite agent gets interesting when you add a super-agent to it, responsible for picking a sub-agent to play a specific game in a boX match. I would imagine the super-agent would then be trained similar to a Texas Hold'em agent and converge to game-theoretical optima for cheese / standard ratio etc.


This actually has a technical term in machine learning community called ensemble learning. But I don't think it is that easy to implement as of yet. And for efficiency sake the single agent is actually very different from a group of agents which will absolutely require quite a bit of parallel processing to achieve (it is not as simple as installing more GPU can solve). And indeed these agents choose to represent the group of all agents in the AlphaStar league will be those encounter many different strategies and still win for the most part overall. It actually is a very difficult problem to introduce "novelty" and still able to adapt mid-game. The current system is simply not having any learning capability on the fly (within one game, in machine learning term, it is a system with offline learning, instead of active/online learning which is much much more difficult).

I don't know anything about AI, but wouldn't it be sufficient to simply have the bots play Bo5's against each other instead of Bo1's during the training phase? Because then they can still learn from what their opponent has been doing in previous games.

Well, you're not wrong. It's just that if you do that and actually want the bot to learn adaption patterns over multiple games, then you need to feed it the previously played games as input.

If you design that mechanism manually, the most simple approach I can think of is to feed it the history of wins/losses together with the respective agent as additional input:

Game 1: Agent "Mass Blinker" - loss
Game 2: Agent "Proxy Gates" - win
...

and so on (the agent names are chosen for illustration only - to the AI it would just be agent 1, agent 2, ...).

But if you don't want to do design anything manually = let the bot self-learn, then you'd have to feed the complete history of entire games as input, which blows up the input quite a bit.


There is already a matchmaking probabilities parameters in the reinforcement learning process as shown in the blog post
[image loading]

In evolutionary algorithm term, it is similar to mating or ranking mechanism put together, if an agent already played against certain agent with certain strategies, it shouldn't "mate" with the same agent as often, but really need to "mate/match" with newly introduced agents/variations, hence novelty search for new blood. However, the probability for playing against losing agents should still be higher I assume, so the policy learning will be able to reward a bit more and certain agents can solidify its sucessful strategies (try a few more times just to make sure so to speak)

yeah that's not what I was talking about. Basically, if you want the decisions to depend on history, you need to feed the history. And if you want pure self-learning you need to feed all of it. I.e. the complete history of the boX match.
50 pts Copper League
imp42
Profile Blog Joined November 2010
398 Posts
January 25 2019 16:28 GMT
#167
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.
50 pts Copper League
Maenander
Profile Joined November 2002
Germany4926 Posts
Last Edited: 2019-01-25 16:38:59
January 25 2019 16:34 GMT
#168
On January 26 2019 01:28 imp42 wrote:
Show nested quote +
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.

I don't agree, the other matchups could present very different challenges.

Also, oversaturate your minerals human noobs! That was embarassing to see, how could we have been so blind. The value is not in the minerals but in the ability to absorb probe losses and to expand at full capacity.
klipik12
Profile Joined March 2012
United States241 Posts
January 25 2019 16:56 GMT
#169
So is this project over, or will it continue?
<(^_^)> || Axiom - CoL - mYi - Prime - ROOT - EG - Acer || WCS Teamleague pls ;-;
jalstar
Profile Blog Joined September 2009
United States8198 Posts
Last Edited: 2019-01-25 17:01:08
January 25 2019 17:00 GMT
#170
On January 26 2019 01:24 imp42 wrote:
Show nested quote +
On January 26 2019 00:51 counting wrote:
On January 26 2019 00:20 imp42 wrote:
On January 25 2019 21:14 Grumbels wrote:
On January 25 2019 13:00 counting wrote:
On January 25 2019 12:45 imp42 wrote:
On January 25 2019 12:14 vesicular wrote:
On January 25 2019 09:40 TheDougler wrote:
You don't know that it was the the camera change that actually was the determining factor here. It could be that Mana had a better idea of what he was up against. It could be that the warp prism threw off the AI's gameplan (I think it's this one). It could be that this AI isn't quite as good as other AIs.

[...]
The final I would say is to play only one agent. Every game used a different agent. It's akin to playing different players. TLO didn't know this when he was playing and played his matches as if it was the same agent and thus tried strats to counter what he just saw in the previous game, which of course didn't work. Playing against a single agent would be quite interesting.

A misconception IMO. There is no conceptual difference between "one agent" and "multiple agents", because you can simply combine x agents into one composite agent (which is exactly what they did).

Compare it to Innovation switching up his macro game with a 3-rax proxy cheese. It's not akin to playing different players, but the same player choosing a different game plan before he starts the game.

The concept of a composite agent gets interesting when you add a super-agent to it, responsible for picking a sub-agent to play a specific game in a boX match. I would imagine the super-agent would then be trained similar to a Texas Hold'em agent and converge to game-theoretical optima for cheese / standard ratio etc.


This actually has a technical term in machine learning community called ensemble learning. But I don't think it is that easy to implement as of yet. And for efficiency sake the single agent is actually very different from a group of agents which will absolutely require quite a bit of parallel processing to achieve (it is not as simple as installing more GPU can solve). And indeed these agents choose to represent the group of all agents in the AlphaStar league will be those encounter many different strategies and still win for the most part overall. It actually is a very difficult problem to introduce "novelty" and still able to adapt mid-game. The current system is simply not having any learning capability on the fly (within one game, in machine learning term, it is a system with offline learning, instead of active/online learning which is much much more difficult).

I don't know anything about AI, but wouldn't it be sufficient to simply have the bots play Bo5's against each other instead of Bo1's during the training phase? Because then they can still learn from what their opponent has been doing in previous games.

Well, you're not wrong. It's just that if you do that and actually want the bot to learn adaption patterns over multiple games, then you need to feed it the previously played games as input.

If you design that mechanism manually, the most simple approach I can think of is to feed it the history of wins/losses together with the respective agent as additional input:

Game 1: Agent "Mass Blinker" - loss
Game 2: Agent "Proxy Gates" - win
...

and so on (the agent names are chosen for illustration only - to the AI it would just be agent 1, agent 2, ...).

But if you don't want to do design anything manually = let the bot self-learn, then you'd have to feed the complete history of entire games as input, which blows up the input quite a bit.


There is already a matchmaking probabilities parameters in the reinforcement learning process as shown in the blog post
[image loading]

In evolutionary algorithm term, it is similar to mating or ranking mechanism put together, if an agent already played against certain agent with certain strategies, it shouldn't "mate" with the same agent as often, but really need to "mate/match" with newly introduced agents/variations, hence novelty search for new blood. However, the probability for playing against losing agents should still be higher I assume, so the policy learning will be able to reward a bit more and certain agents can solidify its sucessful strategies (try a few more times just to make sure so to speak)

yeah that's not what I was talking about. Basically, if you want the decisions to depend on history, you need to feed the history. And if you want pure self-learning you need to feed all of it. I.e. the complete history of the boX match.


I'd bet truly random strategy selection (over the set of good enough strategies) is unbeatable in BoX
Maenander
Profile Joined November 2002
Germany4926 Posts
January 25 2019 17:06 GMT
#171
On January 26 2019 02:00 jalstar wrote:
Show nested quote +
On January 26 2019 01:24 imp42 wrote:
On January 26 2019 00:51 counting wrote:
On January 26 2019 00:20 imp42 wrote:
On January 25 2019 21:14 Grumbels wrote:
On January 25 2019 13:00 counting wrote:
On January 25 2019 12:45 imp42 wrote:
On January 25 2019 12:14 vesicular wrote:
On January 25 2019 09:40 TheDougler wrote:
You don't know that it was the the camera change that actually was the determining factor here. It could be that Mana had a better idea of what he was up against. It could be that the warp prism threw off the AI's gameplan (I think it's this one). It could be that this AI isn't quite as good as other AIs.

[...]
The final I would say is to play only one agent. Every game used a different agent. It's akin to playing different players. TLO didn't know this when he was playing and played his matches as if it was the same agent and thus tried strats to counter what he just saw in the previous game, which of course didn't work. Playing against a single agent would be quite interesting.

A misconception IMO. There is no conceptual difference between "one agent" and "multiple agents", because you can simply combine x agents into one composite agent (which is exactly what they did).

Compare it to Innovation switching up his macro game with a 3-rax proxy cheese. It's not akin to playing different players, but the same player choosing a different game plan before he starts the game.

The concept of a composite agent gets interesting when you add a super-agent to it, responsible for picking a sub-agent to play a specific game in a boX match. I would imagine the super-agent would then be trained similar to a Texas Hold'em agent and converge to game-theoretical optima for cheese / standard ratio etc.


This actually has a technical term in machine learning community called ensemble learning. But I don't think it is that easy to implement as of yet. And for efficiency sake the single agent is actually very different from a group of agents which will absolutely require quite a bit of parallel processing to achieve (it is not as simple as installing more GPU can solve). And indeed these agents choose to represent the group of all agents in the AlphaStar league will be those encounter many different strategies and still win for the most part overall. It actually is a very difficult problem to introduce "novelty" and still able to adapt mid-game. The current system is simply not having any learning capability on the fly (within one game, in machine learning term, it is a system with offline learning, instead of active/online learning which is much much more difficult).

I don't know anything about AI, but wouldn't it be sufficient to simply have the bots play Bo5's against each other instead of Bo1's during the training phase? Because then they can still learn from what their opponent has been doing in previous games.

Well, you're not wrong. It's just that if you do that and actually want the bot to learn adaption patterns over multiple games, then you need to feed it the previously played games as input.

If you design that mechanism manually, the most simple approach I can think of is to feed it the history of wins/losses together with the respective agent as additional input:

Game 1: Agent "Mass Blinker" - loss
Game 2: Agent "Proxy Gates" - win
...

and so on (the agent names are chosen for illustration only - to the AI it would just be agent 1, agent 2, ...).

But if you don't want to do design anything manually = let the bot self-learn, then you'd have to feed the complete history of entire games as input, which blows up the input quite a bit.


There is already a matchmaking probabilities parameters in the reinforcement learning process as shown in the blog post
[image loading]

In evolutionary algorithm term, it is similar to mating or ranking mechanism put together, if an agent already played against certain agent with certain strategies, it shouldn't "mate" with the same agent as often, but really need to "mate/match" with newly introduced agents/variations, hence novelty search for new blood. However, the probability for playing against losing agents should still be higher I assume, so the policy learning will be able to reward a bit more and certain agents can solidify its sucessful strategies (try a few more times just to make sure so to speak)

yeah that's not what I was talking about. Basically, if you want the decisions to depend on history, you need to feed the history. And if you want pure self-learning you need to feed all of it. I.e. the complete history of the boX match.


I'd bet truly random strategy selection (over the set of good enough strategies) is unbeatable in BoX

Especially because the agents don't have specific tendencies like human players. Unlike a human, AlphaStar can switch to a very different agent executing a very different strategy or even switch races without loss of performance.
nimdil
Profile Blog Joined January 2011
Poland3748 Posts
January 25 2019 17:21 GMT
#172
On January 26 2019 01:34 Maenander wrote:
Show nested quote +
On January 26 2019 01:28 imp42 wrote:
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.

I don't agree, the other matchups could present very different challenges.

Also, oversaturate your minerals human noobs! That was embarassing to see, how could we have been so blind. The value is not in the minerals but in the ability to absorb probe losses and to expand at full capacity.

It's interesting whether it applies to humans as well. This tactic might be more applicable to agents with superior unit control and spending regime.
But then again - maybe it's a new meta.
BronzeKnee
Profile Joined March 2011
United States5217 Posts
Last Edited: 2019-01-25 18:08:52
January 25 2019 18:00 GMT
#173
On January 26 2019 01:28 imp42 wrote:
Show nested quote +
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.


6 times? My man... that isn't how this works. That isn't how any of this works. It far greater than 6 times because of the exponential increase of variables.

We used to talk a lot about Starsense. The ability to know something was happening without having any clue it was happening. The AI has no Starsense, clearly. Mana wrecked with a two Immortal drop that crippled it.

The AI is going to always favor aggression for the same reason I do: If you are attacking you are controlling the tempo of the game. That makes it more likely it will win because it reduces the number of variables in an interaction. Defending takes on many more forms.

Attack the AI and it will fall apart. Hold it's attacks (which is harder because defending takes on more variables and it has near perfect micro), and it will fall apart.
Poopi
Profile Blog Joined November 2010
France12790 Posts
January 25 2019 18:23 GMT
#174
I think it’s hard to judge from just one game if it is that vulnerable to opponents attacking.
Sad thing from their AMA thread is that they will stop using this current version of their agent, without even playing more games in camera interface to try to understand its flaws better.
WriterMaru
MyLovelyLurker
Profile Joined April 2007
France756 Posts
January 25 2019 18:33 GMT
#175
On January 26 2019 01:05 BronzeKnee wrote:
I had been a proponent of oversaturating minerals for a long time after I saw Hister do it long ago. I gave up on it later, but I think I will go back to it.

There is so much to learn from an AI that doesn't care about norms or customs and does whatever it thinks gives it the best chance to win. As much as I try to avoid the pressure of following whatever everyone else does, its get me.

Deepmind has a very long way to go to beat Starcraft. A human can regularly take down the cheating SC2 AI with ease, I think Deepmind loses to the cheating AI at this point. It's micro isn't good enough to overcome the massive economic advantage the cheating AI gets.

And why didn't anyone cannon rush the AI? AI is always going to be weak to cheese. It will never play mind games better than a mind.


Lol, my man, Tencent did that in September already, with a bot 'only' circa Platinum lvl, that would get obliterated by AlphaStar : arxiv.org

Respectfully, it might be a good idea to do some research here...
"I just say, it doesn't matter win or lose, I just love Starcraft 2, I love this game, I love this stage, just play like in practice" - TIME/Oliveira
MyLovelyLurker
Profile Joined April 2007
France756 Posts
January 25 2019 18:36 GMT
#176
On January 26 2019 03:23 Poopi wrote:
I think it’s hard to judge from just one game if it is that vulnerable to opponents attacking.
Sad thing from their AMA thread is that they will stop using this current version of their agent, without even playing more games in camera interface to try to understand its flaws better.


Don't worry, they will just look for a more planning oriented approach, and iterate on it. Given the resources they've already invested in this, there's zero chance they take this as a sunk cost, and as they have followed the AlphaGo playbook to the letter so far (with Mana as Fan Hui), this is definitely going all the way to AlphaStar vs (Serral, Maru).
"I just say, it doesn't matter win or lose, I just love Starcraft 2, I love this game, I love this stage, just play like in practice" - TIME/Oliveira
Nakajin
Profile Blog Joined September 2014
Canada8989 Posts
Last Edited: 2019-01-25 18:41:05
January 25 2019 18:38 GMT
#177
On January 26 2019 02:21 nimdil wrote:
Show nested quote +
On January 26 2019 01:34 Maenander wrote:
On January 26 2019 01:28 imp42 wrote:
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.

I don't agree, the other matchups could present very different challenges.

Also, oversaturate your minerals human noobs! That was embarassing to see, how could we have been so blind. The value is not in the minerals but in the ability to absorb probe losses and to expand at full capacity.

It's interesting whether it applies to humans as well. This tactic might be more applicable to agents with superior unit control and spending regime.
But then again - maybe it's a new meta.


I wonder if overbuilding probes is for anticipate probes loss or if it's more something along the line of balancing the risk-reward of expanding.
If you get your expand denied you will have a better incomed if you oversaturate so maybe it's safer against agressive build. Like if you get 1 based and instead of getting the nexus you have 8 more probe in the main you will be ahead in econnomy against the other player witb having only 1 base to defend and when you do get the expand the reward will be imediate so it's less of a risk.
Writerhttp://i.imgur.com/9p6ufcB.jpg
counting
Profile Joined January 2019
11 Posts
January 25 2019 18:44 GMT
#178
On January 26 2019 01:24 imp42 wrote:
Show nested quote +
On January 26 2019 00:51 counting wrote:
On January 26 2019 00:20 imp42 wrote:
On January 25 2019 21:14 Grumbels wrote:
On January 25 2019 13:00 counting wrote:
On January 25 2019 12:45 imp42 wrote:
On January 25 2019 12:14 vesicular wrote:
On January 25 2019 09:40 TheDougler wrote:
You don't know that it was the the camera change that actually was the determining factor here. It could be that Mana had a better idea of what he was up against. It could be that the warp prism threw off the AI's gameplan (I think it's this one). It could be that this AI isn't quite as good as other AIs.

[...]
The final I would say is to play only one agent. Every game used a different agent. It's akin to playing different players. TLO didn't know this when he was playing and played his matches as if it was the same agent and thus tried strats to counter what he just saw in the previous game, which of course didn't work. Playing against a single agent would be quite interesting.

A misconception IMO. There is no conceptual difference between "one agent" and "multiple agents", because you can simply combine x agents into one composite agent (which is exactly what they did).

Compare it to Innovation switching up his macro game with a 3-rax proxy cheese. It's not akin to playing different players, but the same player choosing a different game plan before he starts the game.

The concept of a composite agent gets interesting when you add a super-agent to it, responsible for picking a sub-agent to play a specific game in a boX match. I would imagine the super-agent would then be trained similar to a Texas Hold'em agent and converge to game-theoretical optima for cheese / standard ratio etc.


This actually has a technical term in machine learning community called ensemble learning. But I don't think it is that easy to implement as of yet. And for efficiency sake the single agent is actually very different from a group of agents which will absolutely require quite a bit of parallel processing to achieve (it is not as simple as installing more GPU can solve). And indeed these agents choose to represent the group of all agents in the AlphaStar league will be those encounter many different strategies and still win for the most part overall. It actually is a very difficult problem to introduce "novelty" and still able to adapt mid-game. The current system is simply not having any learning capability on the fly (within one game, in machine learning term, it is a system with offline learning, instead of active/online learning which is much much more difficult).

I don't know anything about AI, but wouldn't it be sufficient to simply have the bots play Bo5's against each other instead of Bo1's during the training phase? Because then they can still learn from what their opponent has been doing in previous games.

Well, you're not wrong. It's just that if you do that and actually want the bot to learn adaption patterns over multiple games, then you need to feed it the previously played games as input.

If you design that mechanism manually, the most simple approach I can think of is to feed it the history of wins/losses together with the respective agent as additional input:

Game 1: Agent "Mass Blinker" - loss
Game 2: Agent "Proxy Gates" - win
...

and so on (the agent names are chosen for illustration only - to the AI it would just be agent 1, agent 2, ...).

But if you don't want to do design anything manually = let the bot self-learn, then you'd have to feed the complete history of entire games as input, which blows up the input quite a bit.


There is already a matchmaking probabilities parameters in the reinforcement learning process as shown in the blog post
[image loading]

In evolutionary algorithm term, it is similar to mating or ranking mechanism put together, if an agent already played against certain agent with certain strategies, it shouldn't "mate" with the same agent as often, but really need to "mate/match" with newly introduced agents/variations, hence novelty search for new blood. However, the probability for playing against losing agents should still be higher I assume, so the policy learning will be able to reward a bit more and certain agents can solidify its sucessful strategies (try a few more times just to make sure so to speak)

yeah that's not what I was talking about. Basically, if you want the decisions to depend on history, you need to feed the history. And if you want pure self-learning you need to feed all of it. I.e. the complete history of the boX match.


I got what you said, there used to be a kind of reinforcement learning distinction in whether to completely recorded all the past actions and behaviors in achieves or always trained from the very beginning independently only treating correlation between scenarios with correlation probability. The problem of utilizing archives is usually about training instability issue, as well as the scaling problem when the combination of possible samples become less representative if we introduce longer and longer training samples to RNNs. (actually input dimention of LSTM or RNNS doesn't have to change, just the training samples become longer, and most likely more hidden layers/nodes for memory)

In essence, LSTM should be able to treat a series of games input as one super long sequences and find patterns across games, hence agents' LSTM initial input will not be initialized when a new game started, but retain the precious iteration's final hidden output as its new initial hidden input. (think of it like the network remembers its final "mental states" after a match and bring it on to the next match). But the first problem will be the supervised learning training examples don't have enough of these kindd of particular player to particular player match-to-match games sequences available, hence the starting the policy networks will not be able to supervised learn them. Hence unless the whole structure is redesigned akin to AlphaGo Zero without any prior supervised learning, or somehow someone painstakingly manually piece together enough high level consecutive training sequence of Go5, Go7 from many high level players in all kinds of different strategies combinations used as training examples.
MyLovelyLurker
Profile Joined April 2007
France756 Posts
January 25 2019 18:44 GMT
#179
On January 26 2019 03:38 Nakajin wrote:
Show nested quote +
On January 26 2019 02:21 nimdil wrote:
On January 26 2019 01:34 Maenander wrote:
On January 26 2019 01:28 imp42 wrote:
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.

I don't agree, the other matchups could present very different challenges.

Also, oversaturate your minerals human noobs! That was embarassing to see, how could we have been so blind. The value is not in the minerals but in the ability to absorb probe losses and to expand at full capacity.

It's interesting whether it applies to humans as well. This tactic might be more applicable to agents with superior unit control and spending regime.
But then again - maybe it's a new meta.


I wonder if overbuilding probes is for anticipate probes loss or if it's more something along the line of balancing the risk-reward of expanding.
If you get your expand denied you will have a better incomed if you oversaturate so maybe it's safer against agressive build. Like if you get 1 based and instead of getting the nexus you have 8 more probe in the main you will be ahead in econnomy against the other player witb having only 1 base to defend and when you do get the expand the reward will be imediate so it's less of a risk.


I think it's the latter. I've gone and looked at minerals lost in the final game where it was sacrificing 2 oracles each time to one-shot Mana's workers. The mineral trade looks even in an instant, but killing workers makes AS better off in the long run, so it takes a very clinical and quantified view of future income. Implicitly embedded in the reinforcement learning algorithm is a very strong Monte-Carlo optimization engine. So once agents are fully trained (in 3 to 6 months IMHO), if they still do whacko stuff like oversaturating minerals, that is probably not a local optimum but the actual thing to copy for us as humans.
"I just say, it doesn't matter win or lose, I just love Starcraft 2, I love this game, I love this stage, just play like in practice" - TIME/Oliveira
imp42
Profile Blog Joined November 2010
398 Posts
January 25 2019 20:00 GMT
#180
On January 26 2019 03:00 BronzeKnee wrote:
Show nested quote +
On January 26 2019 01:28 imp42 wrote:
On January 26 2019 01:22 BronzeKnee wrote:
People keep talking about the limitations that were placed on the AI... how about the limitations placed on the humans?

PvP only?

Pick random. AI has no clue what race you are. Cheese hard or feign cheese and collect the easy win.

From a scientific point of view, expanding from 1 to 6 matchups doesn't add much value. It just costs 6 times more in terms of compute power or time.


6 times? My man... that isn't how this works. That isn't how any of this works. It far greater than 6 times because of the exponential increase of variables.

Well, to expand from 1 mirror matchup to 3 you most definitely just need ~3 times the resources.

I admit in a non-mirror you have roughly twice as may different unit types. But whether that actually increases the number of variables exponentially depends on the implementation of the NN. For example, if you feed raw pixels to a fully connected NN, then it doesn't matter how many unit types you identify on a higher level of abstraction. You will have the same amount of nodes and edges in the neural net.

this is loosely related: https://www.quora.com/How-does-the-training-time-for-a-convolutional-neural-network-scale-with-input-image-size

I am a bit unsure who of us doesn't understand how any of this works at the moment... maybe both
50 pts Copper League
Prev 1 7 8 9 10 11 19 Next All
Please log in or register to reply.
Live Events Refresh
Next event in 1h 7m
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
Creator 119
StarCraft: Brood War
Bisu 2520
firebathero 1366
EffOrt 1055
Shuttle 986
Mini 661
Hyuk 550
Larva 471
Soulkey 318
PianO 295
Soma 243
[ Show more ]
Snow 192
Mind 187
ToSsGirL 146
Rush 135
Pusan 113
Barracks 95
Sharp 78
Hyun 71
Sea.KH 69
JYJ45
Aegong 30
Movie 27
Sacsri 22
yabsab 21
soO 20
Terrorterran 18
JulyZerg 18
GoRush 17
Free 16
HiyA 15
Yoon 13
IntoTheRainbow 13
Bale 12
ivOry 3
Trikslyr0
Dota 2
Gorgc12000
qojqva2303
XcaliburYe230
syndereN229
League of Legends
singsing2345
Dendi1308
Counter-Strike
fl0m870
flusha367
kRYSTAL_167
Other Games
tarik_tv30763
gofns21212
B2W.Neo1405
hiko809
FrodaN609
DeMusliM457
Lowko365
crisheroes360
ArmadaUGS88
QueenE58
Rex26
Organizations
Other Games
gamesdonequick40666
StarCraft 2
angryscii 43
StarCraft: Brood War
Kim Chul Min (afreeca) 10
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 13 non-featured ]
StarCraft 2
• poizon28 11
• AfreecaTV YouTube
• intothetv
• Kozan
• IndyKCrew
• LaughNgamezSOOP
• Migwel
• sooper7s
StarCraft: Brood War
• BSLYoutube
• STPLYoutube
• ZZZeroYoutube
League of Legends
• Nemesis6191
• TFBlade603
Upcoming Events
WardiTV European League
1h 7m
MaNa vs sebesdes
Mixu vs Fjant
ByuN vs HeRoMaRinE
ShoWTimE vs goblin
Gerald vs Babymarine
Krystianer vs YoungYakov
PiGosaur Monday
9h 7m
The PondCast
19h 7m
WardiTV European League
21h 7m
Jumy vs NightPhoenix
Percival vs Nicoract
ArT vs HiGhDrA
MaxPax vs Harstem
Scarlett vs Shameless
SKillous vs uThermal
uThermal 2v2 Circuit
1d 1h
Replay Cast
1d 9h
RSL Revival
1d 19h
ByuN vs SHIN
Clem vs Reynor
Replay Cast
2 days
RSL Revival
2 days
Classic vs Cure
FEL
3 days
[ Show More ]
RSL Revival
3 days
FEL
3 days
FEL
4 days
CSO Cup
4 days
BSL20 Non-Korean Champi…
4 days
Bonyth vs QiaoGege
Dewalt vs Fengzi
Hawk vs Zhanhun
Sziky vs Mihu
Mihu vs QiaoGege
Zhanhun vs Sziky
Fengzi vs Hawk
Sparkling Tuna Cup
4 days
RSL Revival
4 days
FEL
5 days
BSL20 Non-Korean Champi…
5 days
Bonyth vs Dewalt
QiaoGege vs Dewalt
Hawk vs Bonyth
Sziky vs Fengzi
Mihu vs Zhanhun
QiaoGege vs Zhanhun
Fengzi vs Mihu
Replay Cast
6 days
Liquipedia Results

Completed

BSL Season 20
HSC XXVII
Heroes 10 EU

Ongoing

JPL Season 2
BSL 2v2 Season 3
Acropolis #3
KCM Race Survival 2025 Season 2
CSL 17: 2025 SUMMER
Copa Latinoamericana 4
Jiahua Invitational
Championship of Russia 2025
RSL Revival: Season 1
Murky Cup #2
BLAST.tv Austin Major 2025
ESL Impact League Season 7
IEM Dallas 2025
PGL Astana 2025
Asian Champions League '25
BLAST Rivals Spring 2025
MESA Nomadic Masters
CCT Season 2 Global Finals
IEM Melbourne 2025

Upcoming

2025 ACS Season 2: Qualifier
CSLPRO Last Chance 2025
CSL Xiamen Invitational
2025 ACS Season 2
CSLPRO Chat StarLAN 3
K-Championship
uThermal 2v2 Main Event
SEL Season 2 Championship
FEL Cracov 2025
Esports World Cup 2025
Underdog Cup #2
StarSeries Fall 2025
FISSURE Playground #2
BLAST Open Fall 2025
BLAST Open Fall Qual
Esports World Cup 2025
BLAST Bounty Fall 2025
BLAST Bounty Fall Qual
IEM Cologne 2025
FISSURE Playground #1
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2025 TLnet. All Rights Reserved.