• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 18:28
CEST 00:28
KST 07:28
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
[ASL21] Ro16 Preview Pt2: All Star10Team Liquid Map Contest #22 - The Finalists16[ASL21] Ro16 Preview Pt1: Fresh Flow9[ASL21] Ro24 Preview Pt2: News Flash10[ASL21] Ro24 Preview Pt1: New Chaos0
Community News
2026 GSL Season 1 Qualifiers17Maestros of the Game 2 announced92026 GSL Tour plans announced15Weekly Cups (April 6-12): herO doubles, "Villains" prevail1MaNa leaves Team Liquid24
StarCraft 2
General
Maestros of the Game 2 announced 2026 GSL Tour plans announced Team Liquid Map Contest #22 - The Finalists MaNa leaves Team Liquid Blizzard Classic Cup @ BlizzCon 2026 - $100k prize pool
Tourneys
2026 GSL Season 1 Qualifiers INu's Battles#14 <BO.9 2Matches> Sparkling Tuna Cup - Weekly Open Tournament GSL CK: More events planned pending crowdfunding RSL Revival: Season 5 - Qualifiers and Main Event
Strategy
Custom Maps
[D]RTS in all its shapes and glory <3 [A] Nemrods 1/4 players [M] (2) Frigid Storage
External Content
Mutation # 522 Flip My Base The PondCast: SC2 News & Results Mutation # 521 Memorable Boss Mutation # 520 Moving Fees
Brood War
General
BW General Discussion BGH Auto Balance -> http://bghmmr.eu/ FlaSh: This Will Be My Final ASL【ASL S21 Ro.16】 ASL21 General Discussion Data needed
Tourneys
[ASL21] Ro16 Group C [Megathread] Daily Proleagues [ASL21] Ro16 Group D [BSL22] RO16 Tie-Breaker - Sat & Sun 21:00 CEST
Strategy
Simple Questions, Simple Answers What's the deal with APM & what's its true value Any training maps people recommend? Fighting Spirit mining rates
Other Games
General Games
Nintendo Switch Thread Dawn of War IV Diablo IV Total Annihilation Server - TAForever Starcraft Tabletop Miniature Game
Dota 2
The Story of Wings Gaming
League of Legends
G2 just beat GenG in First stand
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Deck construction bug Heroes of StarCraft mini-set
TL Mafia
Vanilla Mini Mafia Mafia Game Mode Feedback/Ideas TL Mafia Community Thread Five o'clock TL Mafia
Community
General
US Politics Mega-thread Things Aren’t Peaceful in Palestine Russo-Ukrainian War Thread Canadian Politics Mega-thread YouTube Thread
Fan Clubs
The IdrA Fan Club
Media & Entertainment
[Manga] One Piece Anime Discussion Thread [Req][Books] Good Fantasy/SciFi books Movie Discussion!
Sports
Formula 1 Discussion 2024 - 2026 Football Thread McBoner: A hockey love story Cricket [SPORT]
World Cup 2022
Tech Support
Strange computer issues (software) [G] How to Block Livestream Ads
TL Community
The Automated Ban List
Blogs
Sexual Health Of Gamers
TrAiDoS
lurker extra damage testi…
StaticNine
Broowar part 2
qwaykee
Funny Nicknames
LUCKY_NOOB
Iranian anarchists: organize…
XenOsky
Customize Sidebar...

Website Feedback

Closed Threads



Active: 1370 users

Deep Learning Starcraft

Forum Index > SC2 General
Post a Reply
tztztz
Profile Joined March 2011
Germany314 Posts
Last Edited: 2016-11-06 23:07:06
November 06 2016 23:04 GMT
#1
As excited as I was about the announcement, as disappointed was I about the fact that they left so many fundamental questions unanswered. They talked about APM limitation and incomplete information and stuff like that, but they didn't talk about the more basic challanges and how they plan to approach them. One of the basic questions I had was: how can you deep learn a game without fixed rules? And how useful is that?

They said an AI that can play Starcraft will need to have the ability to make long-term game plans. But your game plan, at the beginnig of the game as well as during the game, is heavily dependent on the map and version of the game you are playing. One of the most important skills of a human SC player is to look at a new map or a new balance patch and at least have a basic idea of how this will impact a certain match up. But this is impossible for an AI, at least for the type of AI DeepMind developed till now.

All that DeepMind did with AlphaGo was massivly reducing the searchspace of Go with advanced policy and value networks, fed with millions of Go games. AlphaGo did not somehow gain awareness about complex Go concepts or something like that, it just used its deep learning to be very efficient regarding move selection and board evaluation. But its not "intelligent", it's still a nonthinking machine that just gets better at Go.

Now the problem with Starcraft 2 is: Starcraft 2 is not A game, it's many games. It's as many games as there are combinations of versions and maps. So the question is: how do you feed it's networks?

Let's imagine for a second we had millions of replays of the current version of the game, but only on maps like blistering sands and scrap station. If a network would deep learn only from those replays, it would never be able to learn the game to compete with a pro player on the maps of the current map pool.

What might sound like an extreme example is true for every change of the game. If you would fed the network with replays of any variety of versions and maps, it would only learn the most basic things that are true on those past versions and maps, but might become obsolete on future versions/maps.

If you ever plan to beat a pro player in a showmatch, you have to determine the version and map pool many months in advance, than somehow produce enough replays on those maps to train the networks (which might as well be impossible for starcraft), just to trash your AI with the next balance patch after the showmatch, so it's value network would become so significantly inaccurate (if not completly useless), that it will have to deep learn again what game plans and stategies and unit compositions are good/bad.

Of course it could very well be that DeepMind comes up with completlly different things but that's what I'm saying. They didn't talk about anything at all.
CodeWarriors
Profile Joined November 2016
1 Post
November 07 2016 00:04 GMT
#2
I agree with your skepticism of Alpha Go. I would also add that map darkness creates almost infinite possibilities of proxy combinations. Alpha Go's direction is hard to say without its limitations being public knowledge, the question whether it will learn about proxies or use its computer-precise micro still remains. It would be interesting if Alpha Go was used to play against players instead of random PC's.
snakeeyez
Profile Joined May 2011
United States1231 Posts
Last Edited: 2016-11-07 00:06:34
November 07 2016 00:04 GMT
#3
Yeah its going to be interesting to see how it plays out. In my mind if someone adds a new map never seen before the AI would need to be able to function very competently on it based on previous knowledge and predictions just like any pro player would. Its ability has to be able to transfer from map to map.
They would probably have trouble talking about this with specifics because its not built yet. They might not even know the approach yet. How the whole thing works might not be like GO at all. To me a game like civilization 6 would be more ideal than a game where apm and things count for so much
Chocolate
Profile Blog Joined December 2010
United States2350 Posts
November 07 2016 00:14 GMT
#4
There's this concept in medical imaging called "segmentation" where you use heuristics, context learning, etc. to segment medical images into component parts to be able to automatically get metrics on organ component size, abnormalities, etc.

This will be even easier with something like a starcraft map because it doesn't rely on guesswork like with images.

I imagine a good AI for starcraft would have significant levels of preprocessing of input information outside of the "learning" portion. Preprocessing the map would be extremely useful, both in terms of "mapping" it and in learning in specific map-related contexts (e.g. attacking up ramps). By doing so the AI could more easily develop an understanding of map concepts so that upon preprocessing the layout of a new map, it could draw on its "segmentation" and learning within segmented contexts to get a feel more maps it has never seen before.
Chocolate
Profile Blog Joined December 2010
United States2350 Posts
November 07 2016 00:19 GMT
#5
Also I'm sorry for double posting but I'm taking a machine learning/AI project class this next semester and I really hope I'll be able to do something related to starcraft AI. My university in particular has some professors working on AI in the context of imaging and economics, both of which I could perhaps leverage in my project. Really looking forward to see what other people do with this project!
Nakajin
Profile Blog Joined September 2014
Canada8989 Posts
November 07 2016 00:24 GMT
#6
Very interesting to see how they will approach it indeed.

I still think the APM question is gonna be a big one, because if you have perfect control I feel like it will realy reduce the complexity of the game. Lets say Deepmind can worker rush every game and win at the start, at least against 99% of the players, then the recherche for the optimal way of playing his gonna be a lot simpler and way less interesting because the computer can just limit the amount of unknown information.
Writerhttp://i.imgur.com/9p6ufcB.jpg
tztztz
Profile Joined March 2011
Germany314 Posts
Last Edited: 2016-11-07 00:57:23
November 07 2016 00:45 GMT
#7
On November 07 2016 09:14 Chocolate wrote:
There's this concept in medical imaging called "segmentation" where you use heuristics, context learning, etc. to segment medical images into component parts to be able to automatically get metrics on organ component size, abnormalities, etc.

This will be even easier with something like a starcraft map because it doesn't rely on guesswork like with images.

I imagine a good AI for starcraft would have significant levels of preprocessing of input information outside of the "learning" portion. Preprocessing the map would be extremely useful, both in terms of "mapping" it and in learning in specific map-related contexts (e.g. attacking up ramps). By doing so the AI could more easily develop an understanding of map concepts so that upon preprocessing the layout of a new map, it could draw on its "segmentation" and learning within segmented contexts to get a feel more maps it has never seen before.


Yes, a ramp is mostly the same on every map and the AI can reliably rely on his policy network regarding ramp specific choices. But thats it. His values will only be useful regarding choices that are "right" on the maps the networks were trained on, but useless at really map specific choices.

I get the concept of segmentation, but I'm talking about things you can't segment. I'm talking about expansion timings and expansion patterns, pushout timings, those kinds of things, you know. What expansion is hard to take or hard to attack? Timing attack out of 3 base? Is it crucial to deny the opponents 4th base? Most high level game plans and decisions are really map dependent.
Chocolate
Profile Blog Joined December 2010
United States2350 Posts
November 07 2016 01:19 GMT
#8
I'd argue that segmentation could actually be quite helpful in that area. Segmenting an expansion: use choke points, a distance heuristic, etc. to separate it from the rest of the map. Then calculate the time it takes for your slowest combat unit (perhaps depending on time of game) to travel from the edge to a different segmented area e.g. your opponents expansion.

The "goodness" of an expansion will definitely be harder to solve because I think this is prone to overfitting. I'd actually posit, though, that this isn't as big of a problem as you think. In most maps there isn't a huge amount of choice in where to place your expansions until the late game.
alexanderzero
Profile Joined June 2008
United States659 Posts
November 07 2016 09:37 GMT
#9
The issues you bring up are exactly the reasons why Deepmind wants to develop an AI for Starcraft. I'm surprised that you almost seem upset with them for taking on this task.

One of the basic questions I had was: how can you deep learn a game without fixed rules?


A neural network is designed to optimize some output based on input. It has a function that represents how 'good' the output is, and tends to produce those sorts of outputs. With an Atari game, this is score. I'm not sure what you mean when you say that SC2 doesn't have fixed rules.

AlphaGo used two networks. One for guessing the opponents next move, and one for estimating the percent chance of victory for any particular state of the board.

I don't think anyone here can answer your questions about how Deepmind can achieve success in writing AI's to play RTS games. Your best bet would be to read their research papers yourself and guess at how they might use these new inventions to write an AI for Starcraft.
I am a tournament organizazer.
tztztz
Profile Joined March 2011
Germany314 Posts
November 07 2016 16:55 GMT
#10
On November 07 2016 18:37 alexanderzero wrote:
Your best bet would be to read their research papers yourself


I did. Have you?

Like Chess and many other turn based board games with perfekt information, Go is a game with a two dimensional search space: breadth b and depth d. Breadth is the number of possible moves per position, and depth is the game lenght. The search space is given by b^d. For games like Chess and Go this number is too big to bruteforce a value function.

There are two ways to reduce this search space. First, by reducing the depth through position evaluation. This is nothing new and is exactly how it's done in any chess engine (admittedly it is a lot easier to evaluate a position in Chess than in Go because of parameters like material and stuff). The second way is reducing the breadth, and this was the more unique thing about AlphaGo.

Yes, AlphaGo uses two networks. One is called the policy network and its for outputting a probability distribution over the board. After that, it would only branch in the most probable trees and therefore reduce the breadth from all possible moves to only a fraction of that. The network "learned" the values of this probability distribution by a reward function. The outcome of a game (win or lose) would afterwards update the values for every move of that game.

Now the thing with Starcraft is, how do train a network when the rules of the game change?

Lets say we train AI with millions of replays of the current version of the game with the current map pool. If we do, this network will "learn" to build marauders agains roaches or to build overseers agains DTs. This means those "moves" will have resulted in more wins in those situation und therefore have higher probabilities.

And now lets imagine Blizzard releases a balance patch were overseers lose the detector ability and marauders lose there plus damage versus armored. A human player now figures out that overseers are not good against dts anymore and that marauders are weaker against raoches, just by looking at the patch. But a neural network doesn't. It will still build marauder against roaches like before, because it doesn't care about stats and attributes und counters and stuff. In deep learning nothing is scripted, there is no "do this if that".

A neural network doesn't "know" that you have to build detection against invisible units. It doesn't build overseers against DTs because of detection, but because it proved to be successful. The information of the overseer losing the detector ability doesn't mean anything to it. The move "build an Overseer if you spot DTs" still has a high value in its policy network. And it will stay this way until it has lost enough game against DTs. Than the value will get so low that it will choose another move instead.

This is what i mean with no fixed rules. A balance patch is a change of the game. If you change the game the policy network of your AI becomes useless.
tztztz
Profile Joined March 2011
Germany314 Posts
November 07 2016 17:10 GMT
#11
On November 07 2016 10:19 Chocolate wrote:
I'd argue that segmentation could actually be quite helpful in that area. Segmenting an expansion: use choke points, a distance heuristic, etc. to separate it from the rest of the map. Then calculate the time it takes for your slowest combat unit (perhaps depending on time of game) to travel from the edge to a different segmented area e.g. your opponents expansion.

The "goodness" of an expansion will definitely be harder to solve because I think this is prone to overfitting. I'd actually posit, though, that this isn't as big of a problem as you think. In most maps there isn't a huge amount of choice in where to place your expansions until the late game.


Ok I might be really bad at explanig what I mean in general terms so I'm gonna make an extreme example:

Lets say you train your AI with millions of replays only on maps of the current map pool. And after that, you let it play against a human player on an island map where your starting location is an island and every expansion is an island. A human player at least has a basic idea of what to change in his play and how to approach a matchup. But the AI, which was trained for months and would have crushed any player on the maps it learned, is completly hopelessly lost. It wont break down or anything, it will still play, but it wont know whats the right thing to do because it never saw someone build a warp prism out of one base and be successful. The things that it's suppose to do in this situation are all "bad" things for him, things with very low values in it's policy network.

I know this is extreme, but it's just to illustrate whats true for every two maps, because maps are different. No matter how similar they are, they are different.
imp42
Profile Blog Joined November 2010
398 Posts
November 07 2016 18:12 GMT
#12
On November 08 2016 02:10 tztztz wrote:
Show nested quote +
On November 07 2016 10:19 Chocolate wrote:
I'd argue that segmentation could actually be quite helpful in that area. Segmenting an expansion: use choke points, a distance heuristic, etc. to separate it from the rest of the map. Then calculate the time it takes for your slowest combat unit (perhaps depending on time of game) to travel from the edge to a different segmented area e.g. your opponents expansion.

The "goodness" of an expansion will definitely be harder to solve because I think this is prone to overfitting. I'd actually posit, though, that this isn't as big of a problem as you think. In most maps there isn't a huge amount of choice in where to place your expansions until the late game.


Ok I might be really bad at explanig what I mean in general terms so I'm gonna make an extreme example:

Lets say you train your AI with millions of replays only on maps of the current map pool. And after that, you let it play against a human player on an island map where your starting location is an island and every expansion is an island. A human player at least has a basic idea of what to change in his play and how to approach a matchup. But the AI, which was trained for months and would have crushed any player on the maps it learned, is completly hopelessly lost. It wont break down or anything, it will still play, but it wont know whats the right thing to do because it never saw someone build a warp prism out of one base and be successful. The things that it's suppose to do in this situation are all "bad" things for him, things with very low values in it's policy network.

I know this is extreme, but it's just to illustrate whats true for every two maps, because maps are different. No matter how similar they are, they are different.

Basically you are stating the problem of over-fitting an AI to a specific map, aren't you?
Given enough variance in the test maps I would assume an AI to figure out correlations based on rush distance, choke sizes, etc.

Regarding the changing rules: This is one of several reasons why I don't really understand the choice of picking sc2 over BW, except for marketing/publicity reasons.
Also, my advice @Chocolate is to at least check out bwapi before deciding to go with sc2 in an AI term project.
50 pts Copper League
KelvaroN
Profile Joined May 2003
Finland33 Posts
November 07 2016 19:25 GMT
#13
People people, BW players back in the days became pros by just playing Lost Temple, remember? Main skills can be generalized to any starcraft context independent of the map.
imp42
Profile Blog Joined November 2010
398 Posts
November 07 2016 19:31 GMT
#14
On November 08 2016 04:25 KelvaroN wrote:
People people, BW players back in the days became pros by just playing Lost Temple, remember? Main skills can be generalized to any starcraft context independent of the map.


It probably took those pros a while to figure out a new map though, so we should let an AI have some test runs as well
50 pts Copper League
Gere
Profile Joined December 2010
Germany55 Posts
Last Edited: 2016-11-07 20:36:55
November 07 2016 19:58 GMT
#15
@tztztz: What's your point? What do you want? Deepmind wants to work on making progress on one of the hardest challenges in AI and you complain that they don't come up with solutions on the spot? Anyone being sure of solutions after no thought is a fool. You come up with multiple bad ways to approach the problem. Yes, as you say, they won't work and yes, Deepmind will do a completely different way. You misjudge what a good approach with transfer capabilities could be. Let them advance AI with some steps.

I also think Go is an "easy" exercise compared to Starcraft and deep learning will hit some hard limits with Starcraft. But reasonable Starcraft AIs exist and if anyone can make a significant combination with deep learning architectures, it would be Deepmind.
Please log in or register to reply.
Live Events Refresh
The PiG Daily
21:00
Best Games
Rogue vs ByuN
SHIN vs ByuN
Rogue vs ByuN
TBD vs herO
PiGStarcraft713
LiquipediaDiscussion
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
PiGStarcraft664
ProTech132
StarCraft: Brood War
Britney 8091
firebathero 117
Dota 2
capcasts136
League of Legends
Doublelift2217
Heroes of the Storm
Liquid`Hasu450
Other Games
gofns12861
Gorgc7189
summit1g6689
tarik_tv6390
Grubby4291
FrodaN1413
C9.Mang0282
mouzStarbuck228
ArmadaUGS85
ViBE81
ToD50
Mew2King42
PPMD30
Organizations
Other Games
BasetradeTV452
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
[ Show 17 non-featured ]
StarCraft 2
• mYiSmile136
• Adnapsc2 6
• LaughNgamezSOOP
• sooper7s
• AfreecaTV YouTube
• intothetv
• Migwel
• Kozan
• IndyKCrew
StarCraft: Brood War
• FirePhoenix4
• STPLYoutube
• ZZZeroYoutube
• BSLYoutube
Dota 2
• masondota21228
Other Games
• imaqtpie1068
• Scarra811
• Shiphtur264
Upcoming Events
Replay Cast
1h 32m
WardiTV Map Contest Tou…
12h 32m
Classic vs SHIN
MaxPax vs Percival
herO vs Clem
ByuN vs Rogue
Ladder Legends
16h 32m
uThermal 2v2 Circuit
16h 32m
BSL
20h 32m
Sparkling Tuna Cup
1d 11h
WardiTV Map Contest Tou…
1d 12h
Ladder Legends
1d 16h
BSL
1d 20h
CranKy Ducklings
2 days
[ Show More ]
Replay Cast
2 days
Wardi Open
2 days
Afreeca Starleague
2 days
Soma vs hero
Monday Night Weeklies
2 days
Replay Cast
3 days
Replay Cast
3 days
Afreeca Starleague
3 days
Leta vs YSC
Replay Cast
5 days
The PondCast
5 days
KCM Race Survival
5 days
Replay Cast
6 days
Replay Cast
6 days
Escore
6 days
Liquipedia Results

Completed

Proleague 2026-04-23
RSL Revival: Season 4
NationLESS Cup

Ongoing

BSL Season 22
ASL Season 21
CSL 2026 SPRING (S20)
IPSL Spring 2026
KCM Race Survival 2026 Season 2
Escore Tournament S2: W4
StarCraft2 Community Team League 2026 Spring
WardiTV TLMC #16
Nations Cup 2026
IEM Rio 2026
PGL Bucharest 2026
Stake Ranked Episode 1
BLAST Open Spring 2026
ESL Pro League S23 Finals
ESL Pro League S23 Stage 1&2
PGL Cluj-Napoca 2026

Upcoming

Escore Tournament S2: W5
Acropolis #4
BSL 22 Non-Korean Championship
CSLAN 4
Kung Fu Cup 2026 Grand Finals
HSC XXIX
uThermal 2v2 2026 Main Event
Maestros of the Game 2
2026 GSL S2
RSL Revival: Season 5
2026 GSL S1
XSE Pro League 2026
IEM Cologne Major 2026
Stake Ranked Episode 2
CS Asia Championships 2026
IEM Atlanta 2026
Asian Champions League 2026
PGL Astana 2026
BLAST Rivals Spring 2026
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2026 TLnet. All Rights Reserved.