• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 17:55
CEST 23:55
KST 06:55
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
[ASL20] Ro24 Preview Pt2: Take-Off6[ASL20] Ro24 Preview Pt1: Runway132v2 & SC: Evo Complete: Weekend Double Feature4Team Liquid Map Contest #21 - Presented by Monster Energy9uThermal's 2v2 Tour: $15,000 Main Event18
Community News
Weekly Cups (Aug 18-24): herO dethrones MaxPax5Maestros of The Game—$20k event w/ live finals in Paris30Weekly Cups (Aug 11-17): MaxPax triples again!13Weekly Cups (Aug 4-10): MaxPax wins a triple6SC2's Safe House 2 - October 18 & 195
StarCraft 2
General
Weekly Cups (Aug 18-24): herO dethrones MaxPax What mix of new and old maps do you want in the next 1v1 ladder pool? (SC2) : A Eulogy for the Six Pool Geoff 'iNcontroL' Robinson has passed away 2v2 & SC: Evo Complete: Weekend Double Feature
Tourneys
WardiTV Mondays Maestros of The Game—$20k event w/ live finals in Paris RSL: Revival, a new crowdfunded tournament series Sparkling Tuna Cup - Weekly Open Tournament Monday Nights Weeklies
Strategy
Custom Maps
External Content
Mutation # 488 What Goes Around Mutation # 487 Think Fast Mutation # 486 Watch the Skies Mutation # 485 Death from Below
Brood War
General
No Rain in ASL20? BW General Discussion Flash On His 2010 "God" Form, Mind Games, vs JD BGH Auto Balance -> http://bghmmr.eu/ [ASL20] Ro24 Preview Pt2: Take-Off
Tourneys
[ASL20] Ro24 Group E [Megathread] Daily Proleagues [ASL20] Ro24 Group D [ASL20] Ro24 Group B
Strategy
Simple Questions, Simple Answers Fighting Spirit mining rates [G] Mineral Boosting Muta micro map competition
Other Games
General Games
Stormgate/Frost Giant Megathread Nintendo Switch Thread General RTS Discussion Thread Dawn of War IV Path of Exile
Dota 2
Official 'what is Dota anymore' discussion
League of Legends
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Heroes of StarCraft mini-set
TL Mafia
TL Mafia Community Thread Vanilla Mini Mafia
Community
General
Things Aren’t Peaceful in Palestine Russo-Ukrainian War Thread US Politics Mega-thread The year 2050 European Politico-economics QA Mega-thread
Fan Clubs
INnoVation Fan Club SKT1 Classic Fan Club!
Media & Entertainment
Anime Discussion Thread Movie Discussion! [Manga] One Piece [\m/] Heavy Metal Thread
Sports
2024 - 2026 Football Thread TeamLiquid Health and Fitness Initiative For 2023 Formula 1 Discussion
World Cup 2022
Tech Support
High temperatures on bridge(s) Gtx660 graphics card replacement Installation of Windows 10 suck at "just a moment"
TL Community
The Automated Ban List TeamLiquid Team Shirt On Sale
Blogs
Evil Gacha Games and the…
ffswowsucks
Breaking the Meta: Non-Stand…
TrAiDoS
INDEPENDIENTE LA CTM
XenOsky
[Girl blog} My fema…
artosisisthebest
Sharpening the Filtration…
frozenclaw
ASL S20 English Commentary…
namkraft
Customize Sidebar...

Website Feedback

Closed Threads



Active: 3206 users

Winner's Advantage in Grand Finals - Page 2

Blogs > motbob
Post a Reply
Prev 1 2 3 Next All
Cheren
Profile Blog Joined September 2013
United States2911 Posts
March 15 2015 00:13 GMT
#21
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.
+
sertas
Profile Joined April 2012
Sweden887 Posts
March 15 2015 00:16 GMT
#22
you're not taking into account the massive psychological deficit of being down 0-2 in a bo5 compared to being down 0-1 in a bo5 with and without the extra game advantage. Turning a 0-2 is almost impossible while 0-1 is very possible.
motbob
Profile Blog Joined July 2008
United States12546 Posts
Last Edited: 2015-03-15 00:28:22
March 15 2015 00:27 GMT
#23
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.
ModeratorGood content always wins.
FFGenerations
Profile Blog Joined April 2011
7088 Posts
Last Edited: 2015-03-15 00:41:34
March 15 2015 00:41 GMT
#24
i get the impression that a bo5 finals between two teams should be about how good they are against one another in a bo5 finals and NOT about how good they are in a bo5 finals where one team has a 1 game advantage for playing better during the earlier stages of the tourney
Cool BW Music Vid - youtube.com/watch?v=W54nlqJ-Nx8 ~~~~~ ᕤ OYSTERS ᕤ CLAMS ᕤ AND ᕤ CUCKOLDS ᕤ ~~~~~~ ༼ ᕤ◕◡◕ ༽ᕤ PUNCH HIM ༼ ᕤ◕◡◕ ༽ᕤ
GeckoXp
Profile Blog Joined June 2013
Germany2016 Posts
March 15 2015 00:44 GMT
#25
On March 15 2015 09:27 motbob wrote:
Show nested quote +
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.


You do realize you're not flipping a simulated coin, but you're using estimators with assumptions, right?
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 00:50 GMT
#26
On March 15 2015 09:44 GeckoXp wrote:
Show nested quote +
On March 15 2015 09:27 motbob wrote:
On March 15 2015 08:57 itsjustatank wrote:
Your generated Elo predictions based on arbitrary distribution choices resulted in differences that do not seem statistically significant. You do no test to prove that they are statistically significant, you just give the differences in observed percentages.

Null hypothesis: there is no statistically significant difference between starting a double-elimination finals 1-0 versus 0-0
Alternate hypothesis: there is a statistically significant difference between starting a double-elimination finals 1-0 versus 0-0

You have not proven whether or not what you got is noise and whether or not there really is a difference between a 1-0 start and a 0-0 start. You just want one of the two, clearly, and think this is enough to want to make a change.

Your argument is completely non-falsifiable right now. Sure, it may work for the internet, but unless you do that extra work you are pissing in the wind with a cloak of statistics making your advocacy look smart to people who do not know what they are reading.

What is the point of worrying about null/alternate hypotheses, usually? The normal case is this: we sat on the side of the curb all day and observed 200 people passing by. 120 of those people were male. Assuming (liberally) that this has been a completely typical day in terms of the composition of people passing by, can we take our 120/200 number and say that people who walk past the curb are more likely to be male than not? Or was what we saw dictated by random chance? We have to use statistical tests to get a P-value and thereby answer that question and see if we can reject the null.

In Excel, those considerations don't really make any sense because we can just increase the sample size to some absurd number. Imagine I simulate my exercise: I generate a random number and create a cell that returns 1 (for male) 51% of the time and 0 (female) 49% of the time. I then run the test 200 times. The test gives me 54.5%; a test with a 1000 "sample size" gave 52.8%; 10000, 51.5%; 50000, 51.036%. As the sample size gets larger and larger, the value observed converges to the "true" value of 51%.

So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.


You do realize you're not flipping a simulated coin, but you're using estimators with assumptions, right?

From my perspective a tournament is just a series of specifically weighted coin flips.
ModeratorGood content always wins.
GeckoXp
Profile Blog Joined June 2013
Germany2016 Posts
Last Edited: 2015-03-15 00:54:25
March 15 2015 00:52 GMT
#27
Yeah, but you use ELO to determine the skill, which uses rather strong assumptions, which makes stuff complicated. It's not really a fair coin toss or a fair dice throw that way. At least from my point of view. But w/e it's getting late.

edit, the a in toss / throw means single. 8[
itsjustatank
Profile Blog Joined November 2010
Hong Kong9154 Posts
Last Edited: 2015-03-15 01:36:12
March 15 2015 01:31 GMT
#28
you are computing and comparing multiple conditional probabilities based on arbitrary Elo distributions. there are a number of problems with this:
  1. you don't just have an Elo arbitrarily, you maintain one through long-term play within a given population of players playing games that are similar to each other. Elo is not an absolute determination, it is an inference based on prior performance. your probability to win and lose and draw is dependent on that prior performance, and the make-up of the population.

  2. Elo is supposed to be distributed normally because that is the fundamental assumption of player skill in that ratings system. this is compounded by the fact that you do not say how many teams are in the simulations, whether they are a sample from a population or whether they are the population. you also never say how many games they play in each stage. you just say they have a given distribution

  3. the real world does not have infinite sample size or pre-arranged and cherrypicked Elo distributions. in the real world skill also isn't accurately determined by Elo. it is a best-guess estimator and it is pretty shitty in all implementations in ESPORTS right now.

  4. im also fairly certain that you cannot draw in dota, and you cannot draw in most games other than starcraft and fighting games.
given this, we are not denying that there is an observed difference between the two. we are talking about whether that observation is significant. this is very important in the grand scheme of things.

at the point where you even admit this in your OP, there isn't much else to say.

On March 14 2015 17:17 motbob wrote:But the difference between formats seems small enough that, if I were an organizer, I would just keep doing what spectators want (no advantage).


at best you win that in your perfect little infinite compting boxes of imaginary players, it is perhaps a tiny bit better to have 1-0 start in the finals of a double-elimination tournament for the winners bracket player.

if it were significant though, then you would be doing more than just cloaking uncertainties with claims of certainties. you'd have a solid basis to go to every tournament designer and have them unfuck their systems. as it is, you don't.
Photographer"nosotros estamos backamos" - setsuko
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 01:57 GMT
#29
Pretty harsh! Good thing we agree on the real-world conclusions to be drawn from this.
ModeratorGood content always wins.
motbob
Profile Blog Joined July 2008
United States12546 Posts
March 15 2015 02:45 GMT
#30
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.

In the absence of perfect seeding, double elim has obvious advantages if you care about more teams than just the winner. People sometimes talk about the "real finals" in tournaments like the GSL; sometimes the two best players land on one side of the bracket. If that's a problem, double elim fixes it.
ModeratorGood content always wins.
Cascade
Profile Blog Joined March 2006
Australia5405 Posts
Last Edited: 2015-03-15 08:49:07
March 15 2015 08:39 GMT
#31
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.
MysteryMeat1
Profile Blog Joined June 2011
United States3292 Posts
March 15 2015 10:07 GMT
#32
In competitive sports the advantage is that you have to play less games and have an easier time to get to the finals.
"Cause ya know, Style before victory." -The greatest mafia player alive
Lucumo
Profile Joined January 2010
6850 Posts
March 15 2015 11:40 GMT
#33
On March 14 2015 17:17 motbob wrote:
The conclusion I derive from these results is this: if tournament organizers are concerned solely with creating a format where the best team wins, they should have GF with a 1-0 advantage. But the difference between formats seems small enough that, if I were an organizer, I would just keep doing what spectators want (no advantage).

Nope, team from winners' side should need to win one bo3, team from losers' side two. It's not called "double elimination" for nothing.
eonrulz
Profile Blog Joined March 2013
United Kingdom225 Posts
Last Edited: 2015-03-15 13:45:03
March 15 2015 13:34 GMT
#34
On March 15 2015 17:39 Cascade wrote:
Show nested quote +
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.


I actually made exactly the same remark on the LiquidDota version of this blog . Errors and standard deviation are important, regardless of how many toys you run, at the very least so we can see how significant it is.

I'd also be interested in seeing the correlation between say, ELO difference between the top two teams and the top team win rate. You'd definitely expect some correlation, but if its too strongly correlated (or the reverse, I guess), then I'd say that there's a bias there, that you'd have to take into account when dealing with the significance of the results. Or do some reweighting in your monte carlo. I mean, maybe its a small thing, but it'd be nice to see.

Edit: my knowledge of statistics comes from particle physics, where we do some weird stuff that isn't necessarily, rigorously mathematically correct. And our monte carlo samples are often >500k events, and we still worry about statistical uncertainties (not to mention systematics, which might come into play here as part of your ELO definitions). Still want to see the errors, though
Boop!
Liquid`Drone
Profile Joined September 2002
Norway28677 Posts
March 15 2015 18:04 GMT
#35
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.


I'm sorry, I actually completely agree that double elimination shouldn't be used for serious competition. But when I started reading about Instant Runoff, it immediately struck me as a pretty sweet voting system. Why does it suck?
Moderator
Cascade
Profile Blog Joined March 2006
Australia5405 Posts
March 15 2015 21:29 GMT
#36
On March 15 2015 22:34 eonrulz wrote:
Show nested quote +
On March 15 2015 17:39 Cascade wrote:
On March 15 2015 09:27 motbob wrote:
So in this context, an appropriate objection isn't "you didn't do a proper statistical test" because we don't care about inferences and P-values here. We can get the true value, or approach it, just by cranking up the number of simulation runs.

Umm, yeah, you kinda have to do some kind of statistical test, or at least convince us in some way that your numbers are accurate enough so that we feel confident that the differences you quote are more than random noise. We can never get the true value by simulation (infinite accuracy computer simulations with infinite computing time have some practical issues unfortunately. Especially in excel. ), but we can often get close enough with enough computing time. it is incredibly important that you make sure that you actually are putting in enough computing time to get sufficiently accurate numbers out. Did you?

For example, in your first example of 51.6% vs 52.2% from 10k runs. This seems to be close enough to flipping a coin, which will have an error of around 1/sqrt(N), which for 10k runs is 1% relative uncertainty, which is exactly the difference you are seeing. So I think I need some convincing that the differences you are quoting are more than just numerical noise. Let me know if you need help.

Nonetheless, the idea of the simulation is great! I love the approach.


I actually made exactly the same remark on the LiquidDota version of this blog . Errors and standard deviation are important, regardless of how many toys you run, at the very least so we can see how significant it is.

I'd also be interested in seeing the correlation between say, ELO difference between the top two teams and the top team win rate. You'd definitely expect some correlation, but if its too strongly correlated (or the reverse, I guess), then I'd say that there's a bias there, that you'd have to take into account when dealing with the significance of the results. Or do some reweighting in your monte carlo. I mean, maybe its a small thing, but it'd be nice to see.

Edit: my knowledge of statistics comes from particle physics, where we do some weird stuff that isn't necessarily, rigorously mathematically correct. And our monte carlo samples are often >500k events, and we still worry about statistical uncertainties (not to mention systematics, which might come into play here as part of your ELO definitions). Still want to see the errors, though

Ahaha, I'm an (ex) particle physicist myself. :D wrote a minimum bias event generator. Qcd phenomenology essentially.

good to see the particle physics kind of thinking around. exactly what are you doing? (Did do?) You location is Switzerland, so I guess LHC?
itsjustatank
Profile Blog Joined November 2010
Hong Kong9154 Posts
Last Edited: 2015-03-15 21:33:40
March 15 2015 21:32 GMT
#37
On March 16 2015 03:04 Liquid`Drone wrote:
Show nested quote +
On March 15 2015 09:13 Cheren wrote:
Every defense of double elimination I've read is tautological. "Double elimination works because teams that lose twice are eliminated." "Double elimination works because everyone gets a second chance."

It's a system with horrible flaws that isn't used in real sports and needs to get out of esports. It is to tournament formats what Instant Runoff is to voting.


I'm sorry, I actually completely agree that double elimination shouldn't be used for serious competition. But when I started reading about Instant Runoff, it immediately struck me as a pretty sweet voting system. Why does it suck?


IRV does not pick the Condorcet winner. Here, an example from Wikipedia:

IRV uses a process of elimination to assign each voter's ballot to their first choice among a dwindling list of remaining candidates until one candidate receives an outright majority of ballots. It does not comply with the Condorcet criterion. Consider, for example, the following vote count of preferences with three candidates {A,B,C}:

      35 A>B>C
      34 C>B>A
      31 B>C>A

In this case, B is preferred to A by 65 votes to 35, and B is preferred to C by 66 to 34, hence B is strongly preferred to both A and C. B must then win according to the Condorcet criterion. Using the rules of IRV, B is ranked first by the fewest voters and is eliminated, and then C wins with the transferred votes from B.

In cases where there is a Condorcet Winner, and where IRV does not choose it, a majority would by definition prefer the Condorcet Winner to the IRV winner.


STV (single-transferable vote) does a better job.
Photographer"nosotros estamos backamos" - setsuko
deliberate
Profile Joined November 2009
Germany5 Posts
Last Edited: 2015-03-15 21:50:52
March 15 2015 21:48 GMT
#38
Just an additional remark about the statistics in the final set for a double elimination tournament:

Assuming we are running a double elimination bracket where all sets are best-of-threes. In the final game the winner of the winners bracket and the winner of the losers bracket meet. As pointed out earlier, the consistent choice of format would be a BO3, and in the case of the participant from the winners bracket losing, another BO3. A more common choice is the BO5 with an 1:0 advantage for the participant from the winners bracket.

Assuming further, that between the two competitors the chance of one of them winning is constant (like team A has a 60% chance of winning against B for all games), we can calculate the probabilites for the total sets.

The following graph shows the chance of winning the whole set for the team from the winners bracket dependent on their chance of winning the individual matches against the team from the losers bracket. The different curves show a standard BO3 and BO5, as well as the double elimination BO3 and the BO5 with winners bracket advantage.

[image loading]

The first observation is, that the BO5 with 1:0 advantage probability curve is similar to the double elimination BO3 curve, which makes it a viable choice as the final set in terms of consistency.
The second observation is the huge advantage of the team from the winners bracket. Even with a 40% win chance against the team from the losers bracket in the individual matches, the overall chance of winning is still >50%.
itsjustatank
Profile Blog Joined November 2010
Hong Kong9154 Posts
Last Edited: 2015-03-15 22:03:01
March 15 2015 21:58 GMT
#39
Assuming that the chance of winning in a game like Dota is a constant is a very big assumption and one that cannot be made safely unless we are talking about a game that is about to be fixed and intentionally thrown or a card game like blackjack in which strength of hand can be seen and the next cards can be pretty safely predicted.

A team may be more likely to win, but the field of predicting human action is not reducible to numbers currently as much as we would love them to be and try. To pretend that we can is the height of arrogance and to tell others we can is to lie with statistics.

We can talk about likelihoods, but we must qualify that with a lot of uncertainty. If it is not qualified, it is lying.
Photographer"nosotros estamos backamos" - setsuko
micronesia
Profile Blog Joined July 2006
United States24698 Posts
Last Edited: 2015-03-15 22:05:46
March 15 2015 22:05 GMT
#40
While it's true that chance of winning is variable with time and depends on a variety of factors, it is unrealistic to try to model those variations. An example is calculating the odds of getting a 300 if you know your odds of getting a strike in bowling. When you get to frames 8, 9, 10, you most likely will get nervous (which can be exacerbated depending how the people around you react), affecting how you bowl. Of course, you are also getting more physically tired as the game progresses, and the conditions of the lane (oil) are slowly changing. The surface of your bowling ball(s) is also changing over time. On a given throw, any of those effects can have a positive or negative effect on your likelihood of throwing a strike.

You can use a simplified model and say the odds of getting a 300 are 1% if you throw strikes with a consistent success rate of about 68 percent. If you try to argue that the model does not account fully for the other variables described above, you are correct, but the only reasonable thing you can do is say there's not point in doing any calculation, then. Instead, we perform the calculation anyway and just acknowledge what was and was not modeled. It is still interesting to determine that you need a 68% chance of getting a strike to roll a 300 one game in 100.

edit: tank, the edit you made to your post while I was typing seems to already address what I was getting at
ModeratorThere are animal crackers for people and there are people crackers for animals.
Prev 1 2 3 Next All
Please log in or register to reply.
Live Events Refresh
Next event in 2h 6m
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
UpATreeSC 130
ProTech96
JuggernautJason84
CosmosSc2 50
StarCraft: Brood War
NaDa 18
Artosis 0
Dota 2
capcasts258
Counter-Strike
Stewie2K591
flusha237
Super Smash Bros
AZ_Axe5
Heroes of the Storm
Liquid`Hasu460
Other Games
summit1g5192
Grubby2987
shahzam573
Pyrionflax159
ViBE151
C9.Mang0142
ZombieGrub40
Organizations
StarCraft 2
angryscii 25
Other Games
BasetradeTV15
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 18 non-featured ]
StarCraft 2
• musti20045 37
• RyuSc2 23
• IndyKCrew
• Migwel
• sooper7s
• AfreecaTV YouTube
• intothetv
• Kozan
• LaughNgamezSOOP
StarCraft: Brood War
• HerbMon 4
• iopq 3
• STPLYoutube
• ZZZeroYoutube
• BSLYoutube
Dota 2
• masondota22095
League of Legends
• TFBlade780
Counter-Strike
• imaqtpie1133
• Shiphtur199
Upcoming Events
PiGosaur Monday
2h 6m
Afreeca Starleague
12h 6m
hero vs Alone
Royal vs Barracks
Replay Cast
1d 2h
The PondCast
1d 12h
WardiTV Summer Champion…
1d 13h
Replay Cast
2 days
LiuLi Cup
2 days
MaxPax vs TriGGeR
ByuN vs herO
Cure vs Rogue
Classic vs HeRoMaRinE
Cosmonarchy
2 days
OyAji vs Sziky
Sziky vs WolFix
WolFix vs OyAji
BSL Team Wars
2 days
Team Hawk vs Team Dewalt
BSL Team Wars
2 days
Team Hawk vs Team Bonyth
[ Show More ]
SC Evo League
3 days
TaeJa vs Cure
Rogue vs threepoint
ByuN vs Creator
MaNa vs Classic
Maestros of the Game
3 days
ShoWTimE vs Cham
GuMiho vs Ryung
Zoun vs Spirit
Rogue vs MaNa
[BSL 2025] Weekly
3 days
SC Evo League
4 days
Maestros of the Game
4 days
SHIN vs Creator
Astrea vs Lambo
Bunny vs SKillous
HeRoMaRinE vs TriGGeR
BSL Team Wars
4 days
Team Bonyth vs Team Sziky
BSL Team Wars
4 days
Team Dewalt vs Team Sziky
Monday Night Weeklies
5 days
Replay Cast
6 days
Sparkling Tuna Cup
6 days
Liquipedia Results

Completed

CSLAN 3
uThermal 2v2 Main Event
HCC Europe

Ongoing

Copa Latinoamericana 4
BSL 20 Team Wars
KCM Race Survival 2025 Season 3
BSL 21 Qualifiers
ASL Season 20
CSL Season 18: Qualifier 1
Acropolis #4 - TS1
SEL Season 2 Championship
WardiTV Summer 2025
Esports World Cup 2025
BLAST Bounty Fall 2025
BLAST Bounty Fall Qual
IEM Cologne 2025
FISSURE Playground #1
BLAST.tv Austin Major 2025

Upcoming

CSL Season 18: Qualifier 2
CSL 2025 AUTUMN (S18)
LASL Season 20
BSL Season 21
BSL 21 Team A
Chzzk MurlocKing SC1 vs SC2 Cup #2
RSL Revival: Season 2
Maestros of the Game
EC S1
Sisters' Call Cup
IEM Chengdu 2025
PGL Masters Bucharest 2025
Thunderpick World Champ.
MESA Nomadic Masters Fall
CS Asia Championships 2025
Roobet Cup 2025
ESL Pro League S22
StarSeries Fall 2025
FISSURE Playground #2
BLAST Open Fall 2025
BLAST Open Fall Qual
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2025 TLnet. All Rights Reserved.