• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 15:55
CEST 21:55
KST 04:55
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
Code S Season 1 - RO8 Preview4[ASL21] Ro8 Preview Pt2: Progenitors8Code S Season 1 - RO12 Group A: Rogue, Percival, Solar, Zoun13[ASL21] Ro8 Preview Pt1: Inheritors16[ASL21] Ro16 Preview Pt2: All Star10
Community News
Maestros of The Game 2 announcement and schedule !7Weekly Cups (April 27-May 4): Clem takes triple0RSL Revival: Season 5 - Qualifiers and Main Event12Code S Season 1 (2026) - RO12 Results12026 GSL Season 1 Qualifiers25
StarCraft 2
General
Code S Season 1 - RO8 Preview Behind the Blue - Team Liquid History Book Weekly Cups (April 27-May 4): Clem takes triple Blizzard Classic Cup @ BlizzCon 2026 - $100k prize pool Code S Season 1 (2026) - RO12 Results
Tourneys
Maestros of The Game 2 announcement and schedule ! GSL Code S Season 1 (2026) Sea Duckling Open (Global, Bronze-Diamond) RSL Revival: Season 5 - Qualifiers and Main Event Sparkling Tuna Cup - Weekly Open Tournament
Strategy
Custom Maps
[D]RTS in all its shapes and glory <3 [A] Nemrods 1/4 players
External Content
Mutation # 524 Death and Taxes The PondCast: SC2 News & Results Mutation # 523 Firewall Mutation # 522 Flip My Base
Brood War
General
Quality of life changes in BW that you will like ? Tulbo's ASL S21 Ro8 Post-Review Why there arent any 256x256 pro maps? Do we have a pimpest plays list? BGH Auto Balance -> http://bghmmr.eu/
Tourneys
[Megathread] Daily Proleagues Escore Tournament StarCraft Season 2 [ASL21] Ro8 Day 4 Small VOD Thread 2.0
Strategy
Simple Questions, Simple Answers Fighting Spirit mining rates What's the deal with APM & what's its true value Any training maps people recommend?
Other Games
General Games
Stormgate/Frost Giant Megathread Path of Exile Nintendo Switch Thread OutLive 25 (RTS Game) Dawn of War IV
Dota 2
The Story of Wings Gaming
League of Legends
G2 just beat GenG in First stand
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Deck construction bug Heroes of StarCraft mini-set
TL Mafia
Vanilla Mini Mafia Mafia Game Mode Feedback/Ideas TL Mafia Community Thread Five o'clock TL Mafia
Community
General
UK Politics Mega-thread US Politics Mega-thread European Politico-economics QA Mega-thread The Letting Off Steam Thread Canadian Politics Mega-thread
Fan Clubs
The IdrA Fan Club
Media & Entertainment
Anime Discussion Thread [Manga] One Piece [Req][Books] Good Fantasy/SciFi books
Sports
2024 - 2026 Football Thread McBoner: A hockey love story Formula 1 Discussion
World Cup 2022
Tech Support
streaming software Strange computer issues (software) [G] How to Block Livestream Ads
TL Community
The Automated Ban List
Blogs
How EEG Data Can Predict Gam…
TrAiDoS
ramps on octagon
StaticNine
Funny Nicknames
LUCKY_NOOB
Customize Sidebar...

Website Feedback

Closed Threads



Active: 1811 users

Winner's Advantage in Grand Finals - Page 2

Blogs > motbob
Post a Reply
Prev 1 2 3 Next All
Cheren
Profile Blog Joined September 2013
United States2911 Posts
March 15 2015 00:13 GMT
#21
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.
+
sertas
Profile Joined April 2012
Sweden890 Posts
March 15 2015 00:16 GMT
#22
you're not taking into account the massive psychological deficit of being down 0-2 in a bo5 compared to being down 0-1 in a bo5 with and without the extra game advantage. Turning a 0-2 is almost impossible while 0-1 is very possible.
motbob
Profile Blog Joined July 2008
United States12546 Posts
Last Edited: 2015-03-15 00:28:22
March 15 2015 00:27 GMT
#23
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.
ModeratorGood content always wins.
FFGenerations
Profile Blog Joined April 2011
7088 Posts
Last Edited: 2015-03-15 00:41:34
March 15 2015 00:41 GMT
#24
i get the impression that a bo5 finals between two teams should be about how good they are against one another in a bo5 finals and NOT about how good they are in a bo5 finals where one team has a 1 game advantage for playing better during the earlier stages of the tourney
Cool BW Music Vid - youtube.com/watch?v=W54nlqJ-Nx8 ~~~~~ ᕤ OYSTERS ᕤ CLAMS ᕤ AND ᕤ CUCKOLDS ᕤ ~~~~~~ ༼ ᕤ◕◡◕ ༽ᕤ PUNCH HIM ༼ ᕤ◕◡◕ ༽ᕤ
GeckoXp
Profile Blog Joined June 2013
Germany2016 Posts
March 15 2015 00:44 GMT
#25
On March 15 2015 09:27 motbob wrote:
Show nested quote +
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.


You do realize you're not flipping a simulated coin, but you're using estimators with assumptions, right?
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 00:50 GMT
#26
On March 15 2015 09:44 GeckoXp wrote:
Show nested quote +
On March 15 2015 09:27 motbob wrote:
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.


You do realize you're not flipping a simulated coin, but you're using estimators with assumptions, right?

From my perspective a tournament is just a series of specifically weighted coin flips.
ModeratorGood content always wins.
GeckoXp
Profile Blog Joined June 2013
Germany2016 Posts
Last Edited: 2015-03-15 00:54:25
March 15 2015 00:52 GMT
#27
Yeah, but you use ELO to determine the skill, which uses rather strong assumptions, which makes stuff complicated. It's not really a fair coin toss or a fair dice throw that way. At least from my point of view. But w/e it's getting late.

edit, the a in toss / throw means single. 8[
itsjustatank
Profile Blog Joined November 2010
Hong Kong9176 Posts
Last Edited: 2015-03-15 01:36:12
March 15 2015 01:31 GMT
#28
you are computing and comparing multiple conditional probabilities based on arbitrary Elo distributions. there are a number of problems with this:
  1. you don't just have an Elo arbitrarily, you maintain one through long-term play within a given population of players playing games that are similar to each other. Elo is not an absolute determination, it is an inference based on prior performance. your probability to win and lose and draw is dependent on that prior performance, and the make-up of the population.

  2. Elo is supposed to be distributed normally because that is the fundamental assumption of player skill in that ratings system. this is compounded by the fact that you do not say how many teams are in the simulations, whether they are a sample from a population or whether they are the population. you also never say how many games they play in each stage. you just say they have a given distribution

  3. the real world does not have infinite sample size or pre-arranged and cherrypicked Elo distributions. in the real world skill also isn't accurately determined by Elo. it is a best-guess estimator and it is pretty shitty in all implementations in ESPORTS right now.

  4. im also fairly certain that you cannot draw in dota, and you cannot draw in most games other than starcraft and fighting games.
given this, we are not denying that there is an observed difference between the two. we are talking about whether that observation is significant. this is very important in the grand scheme of things.

at the point where you even admit this in your OP, there isn't much else to say.

On March 14 2015 17:17 motbob wrote:But the difference between formats seems small enough that, if I were an organizer, I would just keep doing what spectators want (no advantage).


at best you win that in your perfect little infinite compting boxes of imaginary players, it is perhaps a tiny bit better to have 1-0 start in the finals of a double-elimination tournament for the winners bracket player.

if it were significant though, then you would be doing more than just cloaking uncertainties with claims of certainties. you'd have a solid basis to go to every tournament designer and have them unfuck their systems. as it is, you don't.
Photographer"nosotros estamos backamos" - setsuko
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 01:57 GMT
#29
Pretty harsh! Good thing we agree on the real-world conclusions to be drawn from this.
ModeratorGood content always wins.
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 02:45 GMT
#30
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.

In the absence of perfect seeding, double elim has obvious advantages if you care about more teams than just the winner. People sometimes talk about the "real finals" in tournaments like the GSL; sometimes the two best players land on one side of the bracket. If that's a problem, double elim fixes it.
ModeratorGood content always wins.
Cascade
Profile Blog Joined March 2006
Australia5405 Posts
Last Edited: 2015-03-15 08:49:07
March 15 2015 08:39 GMT
#31
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.
MysteryMeat1
Profile Blog Joined June 2011
United States3292 Posts
March 15 2015 10:07 GMT
#32
In competitive sports the advantage is that you have to play less games and have an easier time to get to the finals.
"Cause ya know, Style before victory." -The greatest mafia player alive
Lucumo
Profile Joined January 2010
6850 Posts
March 15 2015 11:40 GMT
#33
On March 14 2015 17:17 motbob wrote:
The conclusion I derive from these results is this: if tournament organizers are concerned solely with creating a format where the best team wins, they should have GF with a 1-0 advantage. But the difference between formats seems small enough that, if I were an organizer, I would just keep doing what spectators want (no advantage).

Nope, team from winners' side should need to win one bo3, team from losers' side two. It's not called "double elimination" for nothing.
eonrulz
Profile Blog Joined March 2013
United Kingdom225 Posts
Last Edited: 2015-03-15 13:45:03
March 15 2015 13:34 GMT
#34
On March 15 2015 17:39 Cascade wrote:
Show nested quote +
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.


I actually made exactly the same remark on the LiquidDota version of this blog . Errors and standard deviation are important, regardless of how many toys you run, at the very least so we can see how significant it is.

I'd also be interested in seeing the correlation between say, ELO difference between the top two teams and the top team win rate. You'd definitely expect some correlation, but if its too strongly correlated (or the reverse, I guess), then I'd say that there's a bias there, that you'd have to take into account when dealing with the significance of the results. Or do some reweighting in your monte carlo. I mean, maybe its a small thing, but it'd be nice to see.

Edit: my knowledge of statistics comes from particle physics, where we do some weird stuff that isn't necessarily, rigorously mathematically correct. And our monte carlo samples are often >500k events, and we still worry about statistical uncertainties (not to mention systematics, which might come into play here as part of your ELO definitions). Still want to see the errors, though
Boop!
Liquid`Drone
Profile Joined September 2002
Norway28796 Posts
March 15 2015 18:04 GMT
#35
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.


I'm sorry, I actually completely agree that double elimination shouldn't be used for serious competition. But when I started reading about Instant Runoff, it immediately struck me as a pretty sweet voting system. Why does it suck?
Moderator
Cascade
Profile Blog Joined March 2006
Australia5405 Posts
March 15 2015 21:29 GMT
#36
On March 15 2015 22:34 eonrulz wrote:
Show nested quote +
On March 15 2015 17:39 Cascade wrote:
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.


I actually made exactly the same remark on the LiquidDota version of this blog . Errors and standard deviation are important, regardless of how many toys you run, at the very least so we can see how significant it is.

I'd also be interested in seeing the correlation between say, ELO difference between the top two teams and the top team win rate. You'd definitely expect some correlation, but if its too strongly correlated (or the reverse, I guess), then I'd say that there's a bias there, that you'd have to take into account when dealing with the significance of the results. Or do some reweighting in your monte carlo. I mean, maybe its a small thing, but it'd be nice to see.

Edit: my knowledge of statistics comes from particle physics, where we do some weird stuff that isn't necessarily, rigorously mathematically correct. And our monte carlo samples are often >500k events, and we still worry about statistical uncertainties (not to mention systematics, which might come into play here as part of your ELO definitions). Still want to see the errors, though

Ahaha, I'm an (ex) particle physicist myself. :D wrote a minimum bias event generator. Qcd phenomenology essentially.

good to see the particle physics kind of thinking around. exactly what are you doing? (Did do?) You location is Switzerland, so I guess LHC?
itsjustatank
Profile Blog Joined November 2010
Hong Kong9176 Posts
Last Edited: 2015-03-15 21:33:40
March 15 2015 21:32 GMT
#37
On March 16 2015 03:04 Liquid`Drone wrote:
Show nested quote +
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.


I'm sorry, I actually completely agree that double elimination shouldn't be used for serious competition. But when I started reading about Instant Runoff, it immediately struck me as a pretty sweet voting system. Why does it suck?


IRV does not pick the Condorcet winner. Here, an example from Wikipedia:

IRV uses a process of elimination to assign each voter's ballot to their first choice among a dwindling list of remaining candidates until one candidate receives an outright majority of ballots. It does not comply with the Condorcet criterion. Consider, for example, the following vote count of preferences with three candidates {A,B,C}:

      35 A>B>C
      34 C>B>A
      31 B>C>A

In this case, B is preferred to A by 65 votes to 35, and B is preferred to C by 66 to 34, hence B is strongly preferred to both A and C. B must then win according to the Condorcet criterion. Using the rules of IRV, B is ranked first by the fewest voters and is eliminated, and then C wins with the transferred votes from B.

In cases where there is a Condorcet Winner, and where IRV does not choose it, a majority would by definition prefer the Condorcet Winner to the IRV winner.


STV (single-transferable vote) does a better job.
Photographer"nosotros estamos backamos" - setsuko
deliberate
Profile Joined November 2009
Germany5 Posts
Last Edited: 2015-03-15 21:50:52
March 15 2015 21:48 GMT
#38
Just an additional remark about the statistics in the final set for a double elimination tournament:

Assuming we are running a double elimination bracket where all sets are best-of-threes. In the final game the winner of the winners bracket and the winner of the losers bracket meet. As pointed out earlier, the consistent choice of format would be a BO3, and in the case of the participant from the winners bracket losing, another BO3. A more common choice is the BO5 with an 1:0 advantage for the participant from the winners bracket.

Assuming further, that between the two competitors the chance of one of them winning is constant (like team A has a 60% chance of winning against B for all games), we can calculate the probabilites for the total sets.

The following graph shows the chance of winning the whole set for the team from the winners bracket dependent on their chance of winning the individual matches against the team from the losers bracket. The different curves show a standard BO3 and BO5, as well as the double elimination BO3 and the BO5 with winners bracket advantage.

[image loading]

The first observation is, that the BO5 with 1:0 advantage probability curve is similar to the double elimination BO3 curve, which makes it a viable choice as the final set in terms of consistency.
The second observation is the huge advantage of the team from the winners bracket. Even with a 40% win chance against the team from the losers bracket in the individual matches, the overall chance of winning is still >50%.
itsjustatank
Profile Blog Joined November 2010
Hong Kong9176 Posts
Last Edited: 2015-03-15 22:03:01
March 15 2015 21:58 GMT
#39
Assuming that the chance of winning in a game like Dota is a constant is a very big assumption and one that cannot be made safely unless we are talking about a game that is about to be fixed and intentionally thrown or a card game like blackjack in which strength of hand can be seen and the next cards can be pretty safely predicted.

A team may be more likely to win, but the field of predicting human action is not reducible to numbers currently as much as we would love them to be and try. To pretend that we can is the height of arrogance and to tell others we can is to lie with statistics.

We can talk about likelihoods, but we must qualify that with a lot of uncertainty. If it is not qualified, it is lying.
Photographer"nosotros estamos backamos" - setsuko
micronesia
Profile Blog Joined July 2006
United States24772 Posts
Last Edited: 2015-03-15 22:05:46
March 15 2015 22:05 GMT
#40
While it's true that chance of winning is variable with time and depends on a variety of factors, it is unrealistic to try to model those variations. An example is calculating the odds of getting a 300 if you know your odds of getting a strike in bowling. When you get to frames 8, 9, 10, you most likely will get nervous (which can be exacerbated depending how the people around you react), affecting how you bowl. Of course, you are also getting more physically tired as the game progresses, and the conditions of the lane (oil) are slowly changing. The surface of your bowling ball(s) is also changing over time. On a given throw, any of those effects can have a positive or negative effect on your likelihood of throwing a strike.

You can use a simplified model and say the odds of getting a 300 are 1% if you throw strikes with a consistent success rate of about 68 percent. If you try to argue that the model does not account fully for the other variables described above, you are correct, but the only reasonable thing you can do is say there's not point in doing any calculation, then. Instead, we perform the calculation anyway and just acknowledge what was and was not modeled. It is still interesting to determine that you need a 68% chance of getting a strike to roll a 300 one game in 100.

edit: tank, the edit you made to your post while I was typing seems to already address what I was getting at
ModeratorThere are animal crackers for people and there are people crackers for animals.
Prev 1 2 3 Next All
Please log in or register to reply.
Live Events Refresh
Next event in 2h 5m
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
RotterdaM 956
BRAT_OK 127
JuggernautJason111
CosmosSc2 38
StarCraft: Brood War
Britney 16892
EffOrt 777
Mini 572
Dewaltoss 141
ggaemo 135
firebathero 65
Hyun 45
hero 39
Pusan 22
NaDa 10
[ Show more ]
ajuk12(nOOB) 8
Dota 2
XaKoH 423
monkeys_forever344
Counter-Strike
fl0m1790
Fnx 1532
Heroes of the Storm
Liquid`Hasu381
MindelVK9
Other Games
gofns21319
FrodaN2059
Liquid`RaSZi1388
B2W.Neo689
KnowMe483
shahzam409
C9.Mang0245
mouzStarbuck189
ArmadaUGS171
QueenE92
Trikslyr45
ZombieGrub31
Mew2King22
kaitlyn15
Organizations
Other Games
gamesdonequick5899
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
[ Show 17 non-featured ]
StarCraft 2
• Adnapsc2 32
• intothetv
• LaughNgamezSOOP
• AfreecaTV YouTube
• sooper7s
• Migwel
• Kozan
• IndyKCrew
StarCraft: Brood War
• blackmanpl 39
• 80smullet 22
• FirePhoenix7
• STPLYoutube
• ZZZeroYoutube
• BSLYoutube
Dota 2
• lizZardDota273
League of Legends
• imaqtpie2041
Other Games
• Shiphtur235
Upcoming Events
OSC
2h 5m
The PiG Daily
3h 5m
Maru vs Rogue
TBD vs Classic
herO vs Solar
ByuN vs Solar
Replay Cast
4h 5m
CranKy Ducklings
14h 5m
RSL Revival
14h 5m
SHIN vs Bunny
ByuN vs Shameless
WardiTV Invitational
15h 5m
Krystianer vs TriGGeR
Cure vs Rogue
SC Evo League
17h 5m
uThermal 2v2 Circuit
19h 5m
BSL
23h 5m
Artosis vs TerrOr
spx vs StRyKeR
Replay Cast
1d 4h
[ Show More ]
Sparkling Tuna Cup
1d 14h
RSL Revival
1d 14h
Cure vs Zoun
Clem vs Lambo
WardiTV Invitational
1d 15h
BSL
1d 23h
Dewalt vs DragOn
Aether vs Jimin
GSL
2 days
Afreeca Starleague
2 days
Soma vs Leta
Wardi Open
2 days
Monday Night Weeklies
2 days
OSC
3 days
CranKy Ducklings
3 days
Afreeca Starleague
3 days
Light vs Flash
Replay Cast
4 days
Replay Cast
5 days
The PondCast
5 days
Replay Cast
6 days
RSL Revival
6 days
Liquipedia Results

Completed

Proleague 2026-05-07
WardiTV TLMC #16
Nations Cup 2026

Ongoing

BSL Season 22
ASL Season 21
CSL 2026 SPRING (S20)
IPSL Spring 2026
KCM Race Survival 2026 Season 2
Acropolis #4
Proleague 2026-05-08
SCTL 2026 Spring
RSL Revival: Season 5
2026 GSL S1
BLAST Rivals Spring 2026
IEM Rio 2026
PGL Bucharest 2026
Stake Ranked Episode 1
BLAST Open Spring 2026
ESL Pro League S23 Finals
ESL Pro League S23 Stage 1&2

Upcoming

KK 2v2 League Season 1
BSL 22 Non-Korean Championship
YSL S3
Escore Tournament S2: W7
Escore Tournament S2: W8
CSLAN 4
Kung Fu Cup 2026 Grand Finals
HSC XXIX
uThermal 2v2 2026 Main Event
Maestros of the Game 2
2026 GSL S2
BLAST Bounty Summer 2026: Closed Qualifier
Stake Ranked Episode 3
XSE Pro League 2026
IEM Cologne Major 2026
Stake Ranked Episode 2
CS Asia Championships 2026
IEM Atlanta 2026
Asian Champions League 2026
PGL Astana 2026
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2026 TLnet. All Rights Reserved.