• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 08:18
CEST 14:18
KST 21:18
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
uThermal's 2v2 Tour: $15,000 Main Event2Serral wins EWC 202543Tournament Spotlight: FEL Cracow 202510Power Rank - Esports World Cup 202580RSL Season 1 - Final Week9
Community News
Weekly Cups (Jul 28-Aug 3): herO doubles up6LiuLi Cup - August 2025 Tournaments4[BSL 2025] H2 - Team Wars, Weeklies & SB Ladder10EWC 2025 - Replay Pack4Google Play ASL (Season 20) Announced63
StarCraft 2
General
The GOAT ranking of GOAT rankings uThermal's 2v2 Tour: $15,000 Main Event RSL Revival patreon money discussion thread Official Ladder Map Pool Update (April 28, 2025) Weekly Cups (Jul 28-Aug 3): herO doubles up
Tourneys
LiuLi Cup - August 2025 Tournaments WardiTV Mondays RSL Season 2 Qualifier Links and Dates StarCraft Evolution League (SC Evo Biweekly) Global Tourney for College Students in September
Strategy
Custom Maps
External Content
Mutation # 485 Death from Below Mutation # 484 Magnetic Pull Mutation #239 Bad Weather Mutation # 483 Kill Bot Wars
Brood War
General
ASL Season 20 Ro24 Groups StarCraft player reflex TE scores BW General Discussion Google Play ASL (Season 20) Announced BGH Auto Balance -> http://bghmmr.eu/
Tourneys
KCM 2025 Season 3 Small VOD Thread 2.0 [Megathread] Daily Proleagues [ASL20] Online Qualifiers Day 2
Strategy
Fighting Spirit mining rates [G] Mineral Boosting Simple Questions, Simple Answers Muta micro map competition
Other Games
General Games
Stormgate/Frost Giant Megathread Total Annihilation Server - TAForever Nintendo Switch Thread Beyond All Reason [MMORPG] Tree of Savior (Successor of Ragnarok)
Dota 2
Official 'what is Dota anymore' discussion
League of Legends
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Heroes of StarCraft mini-set
TL Mafia
TL Mafia Community Thread Vanilla Mini Mafia
Community
General
Russo-Ukrainian War Thread Things Aren’t Peaceful in Palestine US Politics Mega-thread The Games Industry And ATVI European Politico-economics QA Mega-thread
Fan Clubs
INnoVation Fan Club SKT1 Classic Fan Club!
Media & Entertainment
[\m/] Heavy Metal Thread [Manga] One Piece Movie Discussion! Anime Discussion Thread Korean Music Discussion
Sports
2024 - 2025 Football Thread Formula 1 Discussion TeamLiquid Health and Fitness Initiative For 2023
World Cup 2022
Tech Support
Gtx660 graphics card replacement Installation of Windows 10 suck at "just a moment" Computer Build, Upgrade & Buying Resource Thread
TL Community
TeamLiquid Team Shirt On Sale The Automated Ban List
Blogs
Gaming After Dark: Poor Slee…
TrAiDoS
[Girl blog} My fema…
artosisisthebest
Sharpening the Filtration…
frozenclaw
ASL S20 English Commentary…
namkraft
momentary artworks from des…
tankgirl
from making sc maps to makin…
Husyelt
Customize Sidebar...

Website Feedback

Closed Threads



Active: 671 users

BoxeR: "AlphaGo won't beat humans in StarCraft" - Page 20

Forum Index > SC2 General
568 CommentsPost a Reply
Prev 1 18 19 20 21 22 29 Next All
Mendelfist
Profile Joined September 2010
Sweden356 Posts
March 18 2016 13:49 GMT
#381
On March 18 2016 22:14 necrosexy wrote:
Show nested quote +
On March 18 2016 15:07 Mendelfist wrote:
On March 18 2016 14:16 necrosexy wrote:
On March 18 2016 09:38 Veldril wrote:
I have a feeling that most people are either underestimate the complexity of Go (due to it's not being well-known in the west), overestimate the complexity of Starcraft (due to not understanding heuristic or bias), or not understand how the new AI technology work (due to have not read the Nature's paper yet).

Out of curiosity, how many people here have read or skim through the Nature's paper that describe how Alphago works?

http://www.teamliquid.net/forum/viewpost.php?post_id=25502046

Read page 2 section A of the pdf

You are confusing state space with complexity. What's the state space for throwing a basket ball in real life? That would be utterly impossible to do for an AI, right?

Didn't realize there was an AI that can beat NBA players!

I didn't mention beating NBA players. I said "throwing a basket ball". How large do you think the state space is for throwing a basket ball? How many discrete situations can occur, and what relevance do you think that has for how hard it is to do? I'm trying to tell you that the state space size for Starcraft is a red herring.
necrosexy
Profile Joined March 2011
451 Posts
March 19 2016 01:31 GMT
#382
On March 18 2016 22:49 Mendelfist wrote:
Show nested quote +
On March 18 2016 22:14 necrosexy wrote:
On March 18 2016 15:07 Mendelfist wrote:
On March 18 2016 14:16 necrosexy wrote:
On March 18 2016 09:38 Veldril wrote:
I have a feeling that most people are either underestimate the complexity of Go (due to it's not being well-known in the west), overestimate the complexity of Starcraft (due to not understanding heuristic or bias), or not understand how the new AI technology work (due to have not read the Nature's paper yet).

Out of curiosity, how many people here have read or skim through the Nature's paper that describe how Alphago works?

http://www.teamliquid.net/forum/viewpost.php?post_id=25502046

Read page 2 section A of the pdf

You are confusing state space with complexity. What's the state space for throwing a basket ball in real life? That would be utterly impossible to do for an AI, right?

Didn't realize there was an AI that can beat NBA players!

I didn't mention beating NBA players. I said "throwing a basket ball". How large do you think the state space is for throwing a basket ball? How many discrete situations can occur, and what relevance do you think that has for how hard it is to do? I'm trying to tell you that the state space size for Starcraft is a red herring.

I was joking, because your analogy is terrible (e.g.,the goal is static, complete map information).

Static space is a rough measure of complexity. Of course it's not comprehensive (notice it's merely the first thing discussed in the report I linked), but the disparity between sc and chess/go is absurd -- even if you took a fraction. And bear in mind the estimates were excluding other factors that would've made it even worse!

Mendelfist
Profile Joined September 2010
Sweden356 Posts
March 19 2016 07:56 GMT
#383
On March 19 2016 10:31 necrosexy wrote:
Static space is a rough measure of complexity.

No it isn't. Trying put a number on Starcrafts state space size is ridiculous, as is trying to do it on ANY real world problem. It doesn't tell you anything about how hard it is, because it's for all practical purposes always infinite. Starcraft is more similar to real world problems than Go, which I'm sure is why DeepMind thinks it's an interesting problem. For continuous problems you will have to find another number than state space.
mechengineer123
Profile Joined March 2013
Ukraine711 Posts
March 19 2016 23:55 GMT
#384
They either have no idea what they're talking about or they're deliberately just giving their standard PR answers ("I will do my best! I will be victorious!"). AI would destroy humans without a single doubt. The only interesting question would be how low you could limit the AI's APM before humans stand a chance.
StarStruck
Profile Blog Joined April 2010
25339 Posts
March 20 2016 00:56 GMT
#385
On March 20 2016 08:55 mechengineer123 wrote:
They either have no idea what they're talking about or they're deliberately just giving their standard PR answers ("I will do my best! I will be victorious!"). AI would destroy humans without a single doubt. The only interesting question would be how low you could limit the AI's APM before humans stand a chance.


I know some players who had 900+ apm playing BW and they suck. Lot's of people have tried to make AI's really hard for BW and it was still beatable. I'd like to see one advanced enough to even come close at BW.
Poopi
Profile Blog Joined November 2010
France12886 Posts
March 20 2016 03:45 GMT
#386
On March 20 2016 08:55 mechengineer123 wrote:
They either have no idea what they're talking about or they're deliberately just giving their standard PR answers ("I will do my best! I will be victorious!"). AI would destroy humans without a single doubt. The only interesting question would be how low you could limit the AI's APM before humans stand a chance.

Humans would keep destroying AI without a single doubt.

Very convincing right?
WriterMaru
stapla05
Profile Joined July 2011
Australia67 Posts
Last Edited: 2016-03-20 19:07:20
March 20 2016 07:36 GMT
#387
The AI has advantages and disadvantages versing a human for one its micro, bo and macro will be perfect. So it wont make mistakes. So in turn it will be very effective at whatever builds it does which will result in majority of players would start losing. Maybe not the top level player but a majority of the players will lose to it. As this is very expensive project to code an ai at that level it could accomplish tactic that even human cant perform once it starts to adopt these a tactics there is no hope. You could have put your research into the human brain which could have benefit us more but if you what an ai to win us a gaming while where have large issues around the world i don't know.What happened with deep blue will most possible happen again it will find the solution and make zero chance of an error as all human make errors as that what human are like but that said it was coded by a humans so it depends who codes it. But that said i'm on the wall with this one it could go either way as some of the top player are very intelligent people and i'm sure they have something up there sleeve.
http://www.rts-sanctuary.com/Dawn-Of-War/showuser=96956
Piste
Profile Blog Joined July 2006
6177 Posts
March 21 2016 03:52 GMT
#388
On March 20 2016 16:36 stapla05 wrote:
The AI has advantages and disadvantages versing a human for one its micro, bo and macro will be perfect. So it wont make mistakes. So in turn it will be very effective at whatever builds it does which will result in majority of players would start losing. Maybe not the top level player but a majority of the players will lose to it. As this is very expensive project to code an ai at that level it could accomplish tactic that even human cant perform once it starts to adopt these a tactics there is no hope. You could have put your research into the human brain which could have benefit us more but if you what an ai to win us a gaming while where have large issues around the world i don't know.What happened with deep blue will most possible happen again it will find the solution and make zero chance of an error as all human make errors as that what human are like but that said it was coded by a humans so it depends who codes it. But that said i'm on the wall with this one it could go either way as some of the top player are very intelligent people and i'm sure they have something up there sleeve.

I think you're missing the point on AI developement. The point is making an AI cabaple of learning and adapting, making decisions based on that. After a certain point it can start learning about more complex things other than simple computer games. Theyre not trying to make AI that is concentrating on a single game.
Serendib
Profile Joined May 2011
67 Posts
Last Edited: 2016-03-21 06:30:57
March 21 2016 06:29 GMT
#389
Hey Everyone. I'm Dave Churchill and I organize and run the AIIDE Starcraft AI Competition, and I also wrote UAlbertaBot. I've noticed a lot of misinformation in this thread, so rather than reply to everything individually I decided to take the time to write a detailed history of Starcraft AI Competitions for those who are interested, you can find it here:

http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml

In answer to Boxer's claim: I think it is foolish to say that AI will *never* beat humans at Starcraft, however I feel that this is still quite a few years away. Maybe 5-10 years (unless DeepMind is able to do something miraculous akin to AlphaGo, but that seems unlikely). I also believe that the first to beat expert humans will probably end up heavily abusing micromanagement to do so, so then we will probably enter a philosophical debate about what is 'fair' when it comes to dexterity based games.

Also, most people seem to be confused as to the objective that most of us in the RTS AI field have. Most of us are not really trying to make the best Starcraft bots possible, but instead to come up with new AI algorithms for solving hard problems, and then use Starcraft as a test-bed for those algorithms. We could have much stronger bots if we spent countless hours hard-coding strategies and rules, but that isn't very interesting from a true artificial intelligence point of view.

Thanks for all the discussion, it's great to see so many people interested in the topic!
MyLovelyLurker
Profile Joined April 2007
France756 Posts
March 21 2016 15:51 GMT
#390
Massive amounts of misinformation in this thread indeed, especially when it comes to deep reinforcement learning.

To elaborate on the current state of RTS AI, this recent article is well worth your time :
'RTS AI : Problems and techniques'
richoux.fr

"I just say, it doesn't matter win or lose, I just love Starcraft 2, I love this game, I love this stage, just play like in practice" - TIME/Oliveira
Hotshot
Profile Joined November 2004
Canada184 Posts
March 21 2016 19:59 GMT
#391
Obviously an AI could be coded to easily be better then any human. The question is how hard is it to create that AI.

I feel creating this is more along of the lines 'lots of work', compared to an AI of something like Chess or Go that requires a lot of taught and knowledge.

For example in sc there are builds that you can use on certain maps against certain races. Coding an AI to use those builds would be rather easier. Then you could code in functions that know how to adapt strategy based on certain situations (for example a one base all in, or if someone cannon rushes you are X minute at X spot and at X map there is a way to handle it as efficiently as possible). Then you code in functions knowing when to engage, when to run, what positions to fortify, etc. Then you code functions that understand map. Then you can code micro functions, id imagine probe/drone would be so effective human players would need to send 2 workers to deny 1 from permently harassing, or blink micro would be perfect (even with low'ish apm). Then you code functions that abuse the fact humans can't multitask as much (hitting many different spots at once). You can then even write functions knowing how certain people play, expecting certain strategies, knowing what they struggle with (aka: the ai would never forget). Then you can even write a function that can parse tends of thousands of games and better understand opponents and strategies.... etc.

So overall, I feel a sc2 AI would just take a lot of man hours. Unlike a game like chess or Go which each move exponentially increases the possibilities (so the AI needs to be smart enough to trim out all the obvious bad moves, while in a game like sc2 horribly bad news are much more obvious).

If there was a simple way to use something like c++ to code an AI (hooking into the game and getting the data in a nice clean interface) I am sure more people (like myself) would mess around and build diamond/masters level AI's.
KT_Elwood
Profile Joined July 2015
Germany966 Posts
March 21 2016 20:32 GMT
#392
BoxeR: "AlphaGo won't beat humans in StarCraft"

[image loading]

of course not BoxeR, it can only play Go.
"First he eats our dogs, and then he taxes the penguins... Donald Trump truly is the Donald Trump of our generation. " -DPB
MadMod
Profile Joined May 2011
Norway4 Posts
March 21 2016 22:10 GMT
#393
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.
LetaBot
Profile Blog Joined June 2014
Netherlands557 Posts
March 22 2016 00:58 GMT
#394
On March 22 2016 07:10 MadMod wrote:
Show nested quote +
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.



If by theory based you mean complex algorithms:
My bot uses pathfinding algorithms like A* for things like Wall-building and optimizing mineral gathering. I am currently working on influence maps and MCTS for the strategic level. For bot vs bot a heuristic based approach will still get you further at the moment.

There were some papers about learning from replay, but no top bot that I know of used replay analysis.
If you cannot win with 100 apm, win with 100 cpm.
Hotshot
Profile Joined November 2004
Canada184 Posts
March 22 2016 03:15 GMT
#395
On March 22 2016 09:58 LetaBot wrote:
Show nested quote +
On March 22 2016 07:10 MadMod wrote:
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.



If by theory based you mean complex algorithms:
My bot uses pathfinding algorithms like A* for things like Wall-building and optimizing mineral gathering. I am currently working on influence maps and MCTS for the strategic level. For bot vs bot a heuristic based approach will still get you further at the moment.

There were some papers about learning from replay, but no top bot that I know of used replay analysis.


How are you doing this? All 100% using the editor or using third party tools?
vult
Profile Blog Joined February 2012
United States9400 Posts
Last Edited: 2016-03-22 12:21:55
March 22 2016 12:21 GMT
#396
AlphaGo discussed in The Daily Show with Trevor Noah segment last night (March 21st) --
http://www.cc.com/full-episodes/crzxbs/the-daily-show-with-trevor-noah-march-21--2016---shaka-senghor-season-21-ep-21081 -- second segment.

Starcraft also mentioned.
I used to play random, but for you I play very specifically.
MarlieChurphy
Profile Blog Joined January 2013
United States2063 Posts
Last Edited: 2016-03-22 13:02:05
March 22 2016 12:58 GMT
#397
On March 22 2016 21:21 vult wrote:
AlphaGo discussed in The Daily Show with Trevor Noah segment last night (March 21st) --
http://www.cc.com/full-episodes/crzxbs/the-daily-show-with-trevor-noah-march-21--2016---shaka-senghor-season-21-ep-21081 -- second segment.

Starcraft also mentioned.



Segment starts after first commercial about 10 min in. Not really a mention, just use SC as part of his joke saying he cant even beat computers in it or FIFA
RIP SPOR 11/24/11 NEVAR FORGET
MadMod
Profile Joined May 2011
Norway4 Posts
March 22 2016 20:36 GMT
#398
On March 22 2016 09:58 LetaBot wrote:
Show nested quote +
On March 22 2016 07:10 MadMod wrote:
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.



If by theory based you mean complex algorithms:
My bot uses pathfinding algorithms like A* for things like Wall-building and optimizing mineral gathering. I am currently working on influence maps and MCTS for the strategic level. For bot vs bot a heuristic based approach will still get you further at the moment.

There were some papers about learning from replay, but no top bot that I know of used replay analysis.


That is very interesting. To create a good search space for the MCTS seems extremely hard. It would be awesome to see a very adaptable bot though.

I get the feeling from your answer that the current more adaptable bots play better against humans compared to the less adaptable ones,even though they are not the best in bot vs bot, is his true?
mierin
Profile Joined August 2010
United States4943 Posts
March 22 2016 20:50 GMT
#399
On March 19 2016 16:56 Mendelfist wrote:
Show nested quote +
On March 19 2016 10:31 necrosexy wrote:
Static space is a rough measure of complexity.

For continuous problems you will have to find another number than state space.


I kind of wonder if there is such a thing as a continuous problem...at least for playing games based on thought.
JD, Stork, Calm, Hyuk Fighting!
LetaBot
Profile Blog Joined June 2014
Netherlands557 Posts
March 22 2016 20:55 GMT
#400
On March 22 2016 12:15 Hotshot wrote:
Show nested quote +
On March 22 2016 09:58 LetaBot wrote:
On March 22 2016 07:10 MadMod wrote:
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.



If by theory based you mean complex algorithms:
My bot uses pathfinding algorithms like A* for things like Wall-building and optimizing mineral gathering. I am currently working on influence maps and MCTS for the strategic level. For bot vs bot a heuristic based approach will still get you further at the moment.

There were some papers about learning from replay, but no top bot that I know of used replay analysis.


How are you doing this? All 100% using the editor or using third party tools?



This is for Brood War. I use the (Wiki)Brood War Application Programming Interface





On March 23 2016 05:36 MadMod wrote:
Show nested quote +
On March 22 2016 09:58 LetaBot wrote:
On March 22 2016 07:10 MadMod wrote:
On March 21 2016 15:29 Serendib wrote:
…
http://webdocs.cs.ualberta.ca/~cdavid/starcraftaicomp/history.shtml
…

Thanks for the article, I was hoping someone creating bots would post in this thread.

After reading the article and the paper posted just afterwards I have some questions:
  • I am still wondering how they theory based bots fare versus the more heuristic based ones, are there any bots that use theory based approaches on the strategic level?
  • There were some mention of bots learning from replays, do you know if this was successful?

I could track this down myself, but then I would have to wade through a lot of hard to interpret papers. So I am hoping for an answer here. Thanks again for the article.



If by theory based you mean complex algorithms:
My bot uses pathfinding algorithms like A* for things like Wall-building and optimizing mineral gathering. I am currently working on influence maps and MCTS for the strategic level. For bot vs bot a heuristic based approach will still get you further at the moment.

There were some papers about learning from replay, but no top bot that I know of used replay analysis.


That is very interesting. To create a good search space for the MCTS seems extremely hard. It would be awesome to see a very adaptable bot though.

I get the feeling from your answer that the current more adaptable bots play better against humans compared to the less adaptable ones,even though they are not the best in bot vs bot, is his true?


Yea you need to reduce the search space to get good results with MCTS.

For now, the bots that are capable of executing one strategy partially well have a better chance defeating a human player. But in a Bo5 the more adaptable stands a better chance.

If you cannot win with 100 apm, win with 100 cpm.
Prev 1 18 19 20 21 22 29 Next All
Please log in or register to reply.
Live Events Refresh
LiuLi Cup
11:00
#1
WardiTV526
TKL 225
IntoTheiNu 49
LiquipediaDiscussion
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
Lowko313
TKL 225
Rex 73
trigger 30
StarCraft: Brood War
Britney 41519
actioN 10209
Sea 4185
Rain 3635
ggaemo 2592
Bisu 1514
Mong 724
Larva 559
BeSt 442
Zeus 442
[ Show more ]
Stork 312
Barracks 301
Pusan 191
Mini 186
Soulkey 161
hero 158
ZerO 142
sSak 113
Snow 98
Dewaltoss 97
Soma 77
Sharp 74
TY 61
Shine 48
Killer 47
JYJ39
soO 31
Sacsri 30
Sexy 29
[sc1f]eonzerg 24
Icarus 23
sorry 17
JulyZerg 16
yabsab 15
Bale 11
sas.Sziky 11
IntoTheRainbow 10
scan(afreeca) 9
Terrorterran 1
EffOrt 1
Dota 2
qojqva2048
Gorgc1595
XcaliburYe620
Fuzer 189
boxi98159
Counter-Strike
zeus476
SPUNJ175
kRYSTAL_34
Other Games
gofns15945
singsing2566
B2W.Neo1320
crisheroes461
DeMusliM306
RotterdaM233
XaKoH 194
SortOf165
hiko97
ArmadaUGS41
ZerO(Twitch)13
KnowMe5
Organizations
Other Games
BasetradeTV22
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 15 non-featured ]
StarCraft 2
• StrangeGG 60
• davetesta20
• AfreecaTV YouTube
• intothetv
• Kozan
• IndyKCrew
• LaughNgamezSOOP
• Migwel
• sooper7s
StarCraft: Brood War
• iopq 1
• BSLYoutube
• STPLYoutube
• ZZZeroYoutube
Dota 2
• WagamamaTV436
League of Legends
• Jankos688
Upcoming Events
uThermal 2v2 Circuit
2h 42m
RSL Revival
13h 42m
RSL Revival
21h 42m
SC Evo League
23h 42m
uThermal 2v2 Circuit
1d 2h
CSO Cup
1d 3h
Sparkling Tuna Cup
1d 21h
uThermal 2v2 Circuit
2 days
Wardi Open
2 days
RotterdaM Event
3 days
[ Show More ]
Replay Cast
3 days
RSL Revival
4 days
The PondCast
5 days
Replay Cast
6 days
LiuLi Cup
6 days
Liquipedia Results

Completed

ASL Season 20: Qualifier #2
FEL Cracow 2025
CC Div. A S7

Ongoing

Copa Latinoamericana 4
Jiahua Invitational
BSL 20 Team Wars
KCM Race Survival 2025 Season 3
BSL 21 Qualifiers
uThermal 2v2 Main Event
HCC Europe
BLAST Bounty Fall Qual
IEM Cologne 2025
FISSURE Playground #1
BLAST.tv Austin Major 2025

Upcoming

ASL Season 20
CSLPRO Chat StarLAN 3
BSL Season 21
BSL 21 Team A
RSL Revival: Season 2
Maestros of the Game
SEL Season 2 Championship
WardiTV Summer 2025
Thunderpick World Champ.
MESA Nomadic Masters Fall
CS Asia Championships 2025
Roobet Cup 2025
ESL Pro League S22
StarSeries Fall 2025
FISSURE Playground #2
BLAST Open Fall 2025
BLAST Open Fall Qual
Esports World Cup 2025
BLAST Bounty Fall 2025
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2025 TLnet. All Rights Reserved.