• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 13:13
CET 18:13
KST 02:13
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
[ASL21] Ro24 Preview Pt1: New Chaos0Team Liquid Map Contest #22 - Presented by Monster Energy5ByuL: The Forgotten Master of ZvT30Behind the Blue - Team Liquid History Book19Clem wins HomeStory Cup 289
Community News
Blizzard Classic Cup @ BlizzCon 2026 - $100k prize pool37Weekly Cups (March 9-15): herO, Clem, ByuN win42026 KungFu Cup Announcement6BGE Stara Zagora 2026 cancelled12Blizzard Classic Cup - Tastosis announced as captains18
StarCraft 2
General
Blizzard Classic Cup @ BlizzCon 2026 - $100k prize pool Potential Updates Coming to the SC2 CN Server Weekly Cups (March 2-8): ByuN overcomes PvT block Weekly Cups (August 25-31): Clem's Last Straw? Weekly Cups (March 9-15): herO, Clem, ByuN win
Tourneys
World University TeamLeague (500$+) | Signups Open RSL Season 4 announced for March-April Sparkling Tuna Cup - Weekly Open Tournament WardiTV Team League Season 10 KSL Week 87
Strategy
Custom Maps
Publishing has been re-enabled! [Feb 24th 2026]
External Content
The PondCast: SC2 News & Results Mutation # 517 Distant Threat Mutation # 516 Specter of Death Mutation # 515 Together Forever
Brood War
General
BGH Auto Balance -> http://bghmmr.eu/ JaeDong's form before ASL [ASL21] Ro24 Preview Pt1: New Chaos ASL21 General Discussion Gypsy to Korea
Tourneys
[Megathread] Daily Proleagues [BSL22] Open Qualifiers & Ladder Tours Small VOD Thread 2.0 IPSL Spring 2026 is here!
Strategy
Simple Questions, Simple Answers Soma's 9 hatch build from ASL Game 2 Fighting Spirit mining rates
Other Games
General Games
General RTS Discussion Thread Stormgate/Frost Giant Megathread Nintendo Switch Thread Path of Exile Dawn of War IV
Dota 2
Official 'what is Dota anymore' discussion The Story of Wings Gaming
League of Legends
G2 just beat GenG in First stand
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Deck construction bug Heroes of StarCraft mini-set
TL Mafia
Five o'clock TL Mafia Mafia Game Mode Feedback/Ideas Vanilla Mini Mafia TL Mafia Community Thread
Community
General
US Politics Mega-thread Things Aren’t Peaceful in Palestine YouTube Thread Canadian Politics Mega-thread Russo-Ukrainian War Thread
Fan Clubs
The IdrA Fan Club
Media & Entertainment
Movie Discussion! [Req][Books] Good Fantasy/SciFi books [Manga] One Piece
Sports
2024 - 2026 Football Thread Cricket [SPORT] Formula 1 Discussion Tokyo Olympics 2021 Thread General nutrition recommendations
World Cup 2022
Tech Support
Laptop capable of using Photoshop Lightroom?
TL Community
The Automated Ban List
Blogs
Funny Nicknames
LUCKY_NOOB
Money Laundering In Video Ga…
TrAiDoS
Iranian anarchists: organize…
XenOsky
FS++
Kraekkling
Shocked by a laser…
Spydermine0240
Unintentional protectionism…
Uldridge
ASL S21 English Commentary…
namkraft
Customize Sidebar...

Website Feedback

Closed Threads



Active: 2372 users

GSL Code S Membership statistical analysis - Page 3

Forum Index > StarCraft 2 Tournaments
Post a Reply
Prev 1 2 3 All
Disastorm
Profile Joined January 2008
United States922 Posts
December 10 2010 22:32 GMT
#41
On December 10 2010 16:51 Mip wrote:
Take season 2 for example, We know FruitDealer is amazing because he won Season 1. FruitDealer loses to Foxer in Ro32, that's going to tell me that Foxer is at least skillful enough to beat FruitDealer, which is quite substancial. Then Foxer goes on to lose against NesTea in a nearly dead even match. That's going to hint toward thinking Foxer and NesTea are about at the same level.


Is this actually true though? Do the calculations always assume things like if player a > b and player b > c then player a >c, because I know this isn't the case in most competitive gaming. There are always many cases of rock paper scissors relationships like a > b, b > c, c >a .
"Don't worry so much man. There won't be any more zergs left to QQ. Lots of QQ about TvT is incoming though I bet." - Vrok 9/21/10
See.Blue
Profile Blog Joined October 2008
United States2673 Posts
December 10 2010 22:39 GMT
#42
On December 10 2010 12:34 Mip wrote:
So I've been working on a SC2 player ranking algorithm (see my other post).

So far I've only used the GSL, and I've only included player rankings, no race bias or map bias, or time-based skill evolution (all in progress and will be implemented as my data quantity increases).

Anyway, so I was looking over the list of Code S players and thought to myself that a lot of those players could easily have lost some of their matches and failed to qualify for Code S. So I wanted to see, based on the data, what was the probability of each player actually being in the Top 32.

Here are the results in a Google Spreadsheet

So as you look at that data, bear in mind, this data only obseving the GSL bracket final 64 player wins/losses is all the data in the world on the subject. This makes the algorithm non-ideal for prediction of the top skilled players. But it is ideal for assessing the uncertainty about the point system in actually getting the best players (at least for the top players).

Also bear in mind, this model implicitly assumes that not-qualifying for top 64 and not registering for the tournament are equivalent, which isn't a fair assumption, but there's no data available to fix this. JookToJung gets the raw end of this assumption. He must be very good to qualify all 3 seasons, but the model sees only his losing in the early rounds. This isn't something I like, but I don't have the proper data to correct this problem at this time.

So the table shows a lot of uncertainty about who actually belongs in Code S. There are plenty that could easy have been Code S if things turned out a slightly differently. July is easily Code S caliber, as is Ret, Loner only needed one more set and he'd be S class.

If I had more data on the qualifying rounds, I'm sure that people like JookToJung would look better. I might look into grouping all the players that have 3 or fewer games into one. Because they are hardly estimable with how little data there is on them.

But the higher up on the spreadsheet you go, the results get a lot more accurate since they are based on more games played. There are players that are clearly Top 32, a lot of people are really good, but the uncertainty associated with knowing their skills is fairly high (completely an artifact of not having a lot of data on them). The way the bracket system works, it just doesn't give very good estimates for the people who get knocked out in the first rounds.

Anyway, it is what it is. It should give you an underlying sense on what kind of information is in the data. You don't have to agree with the results, it's just what the data seem to be pointing to (under the constraints of the assumptions I had to make).


Out of curiosity, as a math person, how did you compute the likelihoods?
GeorgeForeman
Profile Joined April 2005
United States1746 Posts
Last Edited: 2010-12-11 00:45:25
December 11 2010 00:44 GMT
#43
On December 11 2010 05:46 Mip wrote:
@GeorgeForeman and confusedcrib I'm glad you paid attention in your intro stats classes, but in Bayesian statistics, you can integrate over the uncertainty in your estimates to obtain a single number that takes into account all of the uncertainty you have in your estimate. We can say with Bayesian statistics that based on our current state of knowledge (priors + data provided) that the probability of Player X actually being Top 32 is Y%.

That you would bring up a t-test for this model immediately puts you at an intro stats level in my brain. Your instinct is correct for that level of stats knowledge, but in this case, it should not be a concern to you. You should think of those percentages in terms of what I described at the end of the paragraph above.

However, to appease you guys, I added a column of Standard Errors. If you are using your intro stats knowledge,however, you will misinterpret them because they mean different things if your data are not from a normal/gaussian distribution.

For a binary outcome, the variance is prob * (1 - prob), and then the standard error is the square root of that, but you have to throw away any thoughts that, for example, 3 standard errors gives you a confidence interval or any nonsense like that that you are taught in intro stats. For example, for NesTea, if you tried to do that, you'd get a confidence interval that included probabilities greater than 1. To do it properly, you'd have to convert to a odds ratio, compute confidence intervals, then convert back to a probability metric.


Kid, I'm a 4th year grad student working on my dissertation in statistics. I've TAUGHT an intro class. If you're going to talk down to someone, at least make sure you know more than they do. Asking for uncertainty estimates only connotes a "t-test" if you're too narrow-minded to consider anything else. As far as I can understand (which is difficult, since you didn't exactly explain it in either of your OPs) you've calculated a posterior distribution for each player's "true skill level". Using the means of these distributions as point estimates you constructed a ranking of them. (This was your previous post.) You've reported standard errors for these, though I'm not sure what those are. Are these numbers the posterior estimates for the standard deviation? Because that's not the same thing as a standard error.

Now, as best as I can tell, you took all of this data and calculated for each player, i, the probability that this player is better than all but at most 31 other players. In other words:

P(S_i>S_j | j is in T and T contains at most 31 elements)

Now, this last thing seems extraordinarily difficult to calculate, given that your estimates for each S_i all come with their own associated variances and that the posterior distribution is dependent upon each of the other. Basically, you've got a p-dimensional normal distribution (where p is the number of players in your data set) with a very confusing-looking covariance matrix. Maybe there's software that makes such a calculation trivial that I'm not aware of, but to me, that looks like a difficult problem. Bravo for taking the time to solve it.

Assuming this is your approach (and again, I'll emphasize that I'm forced to do a lot of inferring because your actual approach is nowhere explained with any degree of clarity), what you end up with are posterior probability estimates. If that is indeed what your spreadsheet is reporting, then I understand why you didn't report the standard deviation, as it's completely determined by the posterior probability estimate.

That said, I'm not sure how useful this second list is. I think the first (where you estimate each player's skill and rank them) does a far better job of not only giving us an idea of who the best players are but also give us an idea of how volatile the estimates are. This "are they REALLY top 32" stuff just muddles the issue IMO. Particularly, it's easy for people to confuse whether someone has a high probability of being top 32 because they're really, really good or whether it's because you've just got a lot of data that tells you to be pretty sure the guy is solid.

Just my $.02. I remember when I took Bayesian a couple of classmates did an analysis of SC1 where they tried to predict winners of matches based on maps, races, and the amount of days the players had since their last game. (I guess this was to measure prep time or something.) It was pretty fun stuff.
like a school bus through a bunch of kids
Mip
Profile Joined June 2010
United States63 Posts
Last Edited: 2010-12-11 08:15:16
December 11 2010 07:29 GMT
#44
I've been hesitant to be too technical in these threads because most of the audience doesn't have a stats background.

The data is a list of names in this format:
Winner Loser
--------------------
Player1 Player2
Player1 Player2
Player2 Player1
Player2 Player3
etc.

The likelihood is the Bradley-Terry model f(x) = exp(skill1)/(exp(skill1)+exp(skill2)).

The priors on the skill parameters are Normal(0,sigma^2) (Bradley Terry model is only dependent on the difference of the skills. Players with skills 100 and 101 would yield the same probability comparisons as if we subtracted 100 to make it 0 and 1, so the 0 mean is arbitrary. It's has same theoretical backing that the ELO system is based off of)

My professor said that sigma^2 could probably be fixed, to test, I just gave it a somewhat informative prior around 1 to see if it the data would alter it (they did not).

So the parameters are run through an MCMC algorithm. Had to use Metropolis steps to calculate draws from the posterior distributions of the skill parameters.

My first report was the mean of the posterior draws and the standard deviation of the posterior draws, then the mean - 2 standard deviations to give a sort of, "at their worst" skill parameter.

The second report, I took each draw from the skill parameters and took the top 32 for each one. Then I calculated the proportion of the times each player appeared in the top 32 over all posterior draws.
Vorlik
Profile Joined October 2010
1522 Posts
December 11 2010 08:02 GMT
#45
This is fascinating. I like it! :-]
Prev 1 2 3 All
Please log in or register to reply.
Live Events Refresh
LAN Event
16:00
StarCraft Madness Day 2
Airneanach58
Liquipedia
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
LamboSC2 311
MindelVK 36
EmSc Tv 17
Codebar 7
StarCraft: Brood War
Britney 27673
Calm 3306
Shuttle 381
Mini 357
BeSt 305
Light 200
EffOrt 193
hero 179
firebathero 164
actioN 108
[ Show more ]
Rush 101
Dewaltoss 73
ggaemo 57
Shine 35
Nal_rA 26
Free 25
IntoTheRainbow 24
GoRush 22
JulyZerg 21
ivOry 5
Dota 2
Gorgc7331
League of Legends
Reynor56
Counter-Strike
fl0m4535
Fnx 3024
byalli500
Super Smash Bros
hungrybox661
Heroes of the Storm
Khaldor527
Liquid`Hasu478
Other Games
singsing2225
Grubby2033
FrodaN1637
Liquid`RaSZi1302
B2W.Neo1008
Beastyqt460
Mlord359
Hui .137
Organizations
Other Games
gamesdonequick624
Dota 2
PGL Dota 2 - Main Stream53
StarCraft 2
EmSc Tv 17
EmSc2Tv 17
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 18 non-featured ]
StarCraft 2
• LUISG 26
• Reevou 4
• LaughNgamezSOOP
• sooper7s
• AfreecaTV YouTube
• Migwel
• intothetv
• Kozan
• IndyKCrew
StarCraft: Brood War
• HerbMon 17
• STPLYoutube
• ZZZeroYoutube
• BSLYoutube
Dota 2
• lizZardDota254
League of Legends
• Nemesis2884
• Jankos1943
• Shiphtur302
Other Games
• imaqtpie280
Upcoming Events
BSL
2h 47m
Replay Cast
15h 47m
Afreeca Starleague
16h 47m
Sharp vs Scan
Rain vs Mong
Wardi Open
18h 47m
Monday Night Weeklies
23h 47m
Sparkling Tuna Cup
1d 16h
Afreeca Starleague
1d 16h
Soulkey vs Ample
JyJ vs sSak
Replay Cast
2 days
Afreeca Starleague
2 days
hero vs YSC
Larva vs Shine
Kung Fu Cup
2 days
[ Show More ]
Replay Cast
3 days
KCM Race Survival
3 days
The PondCast
3 days
WardiTV Team League
3 days
Replay Cast
4 days
WardiTV Team League
4 days
RSL Revival
5 days
Cure vs Zoun
herO vs Rogue
WardiTV Team League
5 days
Platinum Heroes Events
5 days
BSL
6 days
RSL Revival
6 days
ByuN vs Maru
MaxPax vs TriGGeR
WardiTV Team League
6 days
Liquipedia Results

Completed

Jeongseon Sooper Cup
WardiTV Winter 2026
Underdog Cup #3

Ongoing

KCM Race Survival 2026 Season 1
BSL Season 22
CSL Elite League 2026
CSL Season 20: Qualifier 1
RSL Revival: Season 4
Nations Cup 2026
NationLESS Cup
BLAST Open Spring 2026
ESL Pro League S23 Finals
ESL Pro League S23 Stage 1&2
PGL Cluj-Napoca 2026
IEM Kraków 2026
BLAST Bounty Winter 2026
BLAST Bounty Winter Qual

Upcoming

ASL Season 21
Acropolis #4 - TS6
2026 Changsha Offline CUP
CSL Season 20: Qualifier 2
CSL 2026 SPRING (S20)
Acropolis #4
IPSL Spring 2026
Kung Fu Cup 2026 Grand Finals
HSC XXIX
uThermal 2v2 2026 Main Event
IEM Cologne Major 2026
Stake Ranked Episode 2
CS Asia Championships 2026
Asian Champions League 2026
IEM Atlanta 2026
PGL Astana 2026
BLAST Rivals Spring 2026
CCT Season 3 Global Finals
IEM Rio 2026
PGL Bucharest 2026
Stake Ranked Episode 1
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2026 TLnet. All Rights Reserved.