My favorite example is Monty Hall problem. "Smart" people often use it as the evidence of how bad general people are at probability.
It really isn't. The problem is usually given in this form:
> Suppose you're on a game show, and you're given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what's behind the doors, opens another door, say No. 3, which has a goat. He then says to you, "Do you want to pick door No. 2?" Is it to your advantage to switch your choice? [1]
The only correct ansewr to this question is "I don't know." People who answer yes are just taking way too many standardized tests.
The difference here is that the Monty Hall problem has an explanation that while counterintuitive, is statistically sound. You should always switch, because the probability you picked the correct door is locked in at the time you made the choice between 3 doors. It is 1/3 that you picked correctly, and 2/3 that you picked incorrectly. The counterintuitive part is that if you switch, you are effectively selecting all the doors you did not pick originally. It's your original door, vs the field.
The green hat problem hinges on subjective interpretations of the meaning of both "liar" and the different ways in which the liar's sentence may be false. It may be false because the liar owns many hats, none of which are green. Or they own many hats, only some of which are green. Or they own no hats. These are all reasonable interpretations of how the sentence might be false, and the answers presented are not necessarily mutually exclusive.
The problem here is that the usual explanation sneaks in multiple rarely stated assumptions.
If Monty knows the door with the prize and is aiming for the game to continue, then you should switch. (This is the usual argument.)
If Monty doesn't know where the prize is, then you learned nothing. (Monty's result was luck, and he can't impart information that he doesn't have.)
If Monty knows where the prize is and wants you to lose, absolutely don't switch. (Monty will only drag the game out as a way to try to make you lose.)
The reasoning behind these statements is completely solid, and there are no hidden assumptions being snuck in.
I think, canonically, Monty always knows where the prize is, and will always eliminate all doors except for one, and will never eliminate a door with the prize, and will always give you a choice to switch. There's no room for Monty's motivation.
the assumption is told to you in that the game is expected to continue after monty opens a door. if he could open the prize door, the game would describe what happens when he does that
No. That condition holds in all three of the scenarios that I stated. And yet you have 3 different answers. Switching 2/3 win, 1/2 win, and 0/1 win respectively.
You specifically need information that we haven't been given about the counterfactuals. What might Monty have done in other scenarios that we have not yet observed? We don't know. And we're not actually told. That makes that an implicit assumption that wasn't specified.
They're not. It says he "opens another door, say No. 3, which has a goat". That could mean he deliberately chose a door with a goat or he chose one by some unstated process and it happened to have a goat by chance. It says he "knows what's behind the doors" but that statement means nothing because it doesn't say how he uses that knowledge, if at all. It's full of language ambiguity.
That's fine, but given the parameters of the problem change with the Monty Crawl variant, it's not the same problem, and doesn't invalidate the answer of the base variant.
The problem is that the parameter was unspecified in the original problem. Your answer is equivalent to the shaky answer referred to in the paper. Without knowing that the host selects between 2 goats with 50/50 chance, you cannot give a general answer.
The host might be selecting between two goats, or might be selecting between a goat and a car. Either way, it doesn't matter because we don't get any additional information about whether our original choice was correct. (To clarify, this applies to the original problem, not the Crawl variant, where we either sometimes get definitive information, or sometimes get no information)
Edit: Furthermore, I don't think that author's solution to the Crawl problem is correct. When the host eliminates a door, either you will get information that says you should switch, and you'll win 100% of the time; or you won't get information, and you should still always switch and win 2/3 of the time.
> it doesn't matter because we don't get any additional information about whether our original choice was correct
That's the missing assumption. I would say assuming that people are perfectly random falls into the "standardized test" category.
> you won't get information, and you should still always switch and win 2/3 of the time.
You always get some information, the set of possible results becomes narrower, so saying the probabilities don't change is not sufficient. Not a good idea to discuss the problem in informal language though.
Years ago, the Monty Fall variation was mentioned on a local telnet forum I visited. The "consensus" of the thread was the solution of Monty Fall is the same as that of Monty Hall: switching gives you 2/3 chance to win.
That was when I realized many people just memorize the Monty Hall's solution without understanding it, a.k.a. "standardized tests".
I guess I don't understand why you think "I don't know" is the only correct answer. It's clearly not, which is the point of the problem in the first place. It's a bit hard to grok, but once you do, it's clear what the right answer is.
"I don't know" is the only correct anwser because the original problem isn't properly worded.
You need to (at least) add this statement to the original question:
> The host must open a door in any situation, and you know this rule.
Because "host" in daily language is a human being with agency to choose whether to not open a door. Without this statement, the original problem is actually a game theory problem with two players. The host might be playing a strategy where he only opens the door when he knows you picked the car at the first place, or any other strategy.
The above statement removes the ambiguity and that's what the original author meant (but failed to put in words).
(If you're interesting in this, the wikipedia page linked above actually contains a quite extensive discussion in history over it)
I don't think improper wording makes it an example of the type of puzzle GP is complaining about, because the counter-intuitive aspect holds up when everything is clarified.
Ah I see what you mean. Yeah, that makes sense. If the host behavior is not clearly stated then the right answer would indeed be "it can't be determined from available information".
Perhaps there's another formulation you meant to write, but every time I've seen it, it's been equivalent to the thing you just asked, and the only potentially ambiguous part of the question is "to your advantage". If you stay then you have a 1/3 chance of getting a car, and if you switch you have a 2/3 chance.
Yes, it's also correct that you don't "know" the result, and you might prefer goats to cars (even then you should probably sell the car and buy several goats), but there's a reasonable enough interpretation of "advantage" that you shouldn't dismiss the problem outright.
> The version of the Monty Hall problem published in Parade in 1990 did not specifically state that the host would always open another door, or always offer a choice to switch, or even never open the door revealing the car. However, Savant made it clear in her second follow-up column that the intended host's behavior could only be what led to the 2/3 probability she gave as her original answer. (emphasis mine)
My point was that when people ask this question, they often word it like the very 1990 version did, lefting out this critical statement (which Savant considered needed as well, therefore she clarified in the follow-up column), making the question ambiguous.
(Although Savant also said "Very few (out of people who said is 2/3 is wrong) raised questions about ambiguity"... so perhaps people are actually just bad at probablity...?)
Just be glad that you didn't go for the 2 envelopes problem which was her next. Vos Savant was wrong about that one, and few people like the real answer.
The problem is that you're presented with two envelopes with 2 real numbers inside. You randomly select one, then look, and try to guess if you got the larger number. It doesn't seem like you can do better than even, but you can!
Unfortunately everyone hates the answer. Which is that you make up a random number and pretend it is the other one. Your odds of being right are
50% + (probability of choosing between the numbers) / 2
Which can always be strictly bigger than 50%. (Though possibly by only a little amount.)
Special case. If those numbers and yours were all independently randomly chosen from the same distribution, you'll be right 2/3 of the time.
The two reals are selected via some distribution, and the only way you can do better than chance is if you have some knowledge of that distribution.
The question leaves that distribution completely hidden, and your answer smuggles it back in. That feels less like a counter-intuitive math/stats question and more like a badly worded gotcha.
You choose two unequal real numbers. I don't know what they are, and I don't know the distribution from which you choose them. You write them down and put them in separate envelopes.
I'm allowed to choose one envelope and open it to see the number inside, and my job is then to say which envelope holds the larger number.
I claim I have a strategy now which lets me win strictly more than 50% of the time. My strategy is this.
I choose a real number R at random from a distribution that has dense support. In other words, for any two reals, L and U with L<U, P(L<R<U) > 0. This is easy to do ... one method is to list the rationals, positive and negative, then roll a die, discarding numbers until you get a 6.
Now I flip a coin and thereby choose an envelope at random. I proceed by assuming my chosen number is between your two numbers. There is a non-zero chance this is true ... call it e. So e>0.
If I'm wrong then my choice is 50% ... probability is 1-e.
If I'm right then my choice is 100%. ... probability is e.
Combined, my chance of being right is 0.5(1-e) + e = 0.5+e/2, which is strictly greater than 50%.
You can make it as small as you like, and if we play the game repeatedly then you can make it approach 50%. But as it stands, with a one-off game, I can win with a probability that depends on your chosen numbers, but which is strictly bigger than 50%.
I'm not a mathematician so please bear with me here, but I think a problem stems from the fact that the set of reals is "infinite". So, whatever interval you choose, there are infinitely more reals outside the interval as inside (by that I mean that you can fit an infinite number of copies of that interval up to infinity). So the probability e is not >0, it is effectively 0. The second problem is, what does it mean to choose a real at random ? There is an implication that you can choose such number, but as a human living in the finite universe there are limitations to your choice. Any number you can write using all the atoms in the universe is infinitely outnumbered by all numbers that you can't. So effectively it is impossible to pick a random real number. You have to pick a real in some interval, implicitly the interval of reals you can write in an envelope. Which is a different problem than stated originally and for which your "e" can be >0.
> The second problem is, what does it mean to choose a real at random ?... So effectively it is impossible to pick a random real number
Yes, it's established there isn't "uniform distribution over all real numbers" without violating axiom of probability. You're 100% correct on this.
But it doesn't make Colin's solution wrong, because e > 0 for any* well-defined distribution.
> Which is a different problem than stated originally
There are two ways to inteprete the original problem:
A. The numbers are truly randomly picked over all real numbers.
B. The numbers are picked from a well-defined distribution which is unknown to the player.
Since A. is invalid mathematically speaking (without changing the commonly accepted definition of probability), it's reasonable to only consider B., in which case, Colin's solution is correct.
I'm not a mathematician so please bear with me here
I am a mathematician, so please bear with me when I try to explain how this can work.
The rational numbers are countable, and that means that I can write a list of them. There are several ways of doing this, but personally I like the Calkin-Wilf tree[0]. That only gives the positive ones, but we can include zero and the negative ones by interleaving them.
So, whatever interval you choose, there are infinitely more reals outside the interval as inside (by that I mean that you can fit an infinite number of copies of that interval up to infinity). So the probability e is not >0, it is effectively 0.
One you have chosen the two numbers, L and U, I note that there are rational numbers in between. Choose one of those numbers, call it M.
M is in my list above. Now I roll a die, discarding numbers from the list until I get a 6. There is a non-zero probability that the number retained is M, so there is a non-zero probability that my chosen number is between L and U. So e is definitely non-zero.
The second problem is, what does it mean to choose a real at random?
It doesn't have to be uniformly at random -- that's the mistake nearly everyone makes -- and the above process does it perfectly well. It only ever chooses a rational number, but that's OK. It's still a real number, it's still a random number, and for any non-empty interval, there is a non-zero chance the chosen number is inside.
... as a human living in the finite universe there are limitations to your choice.
Yes, but that is accounted for in the explicit description of how to choose the number.
Any number you can write using all the atoms in the universe is infinitely outnumbered by all numbers that you can't.
Again, this is accounted for by the fact that we are not choosing uniformly at random.
Am I right to assume one could also sample from a Gaussian distribution for the method to work? Of course, the probability e of sampling between the two real numbers would be very small, but it would be nonzero.
Yes, any distribution with dense support works. Your problem is to prove that your method of number selection really does have a non-zero probability of being between any two distinct reals.
So ... yes, choose from a Gaussian, but then you have to tell me exactly how.
Yes, this two real numbers question suffers from hidden conditions as the original version of Monty Hall problem does, but even more explicitly.
We can't have an uniform distribution over all real numbers, so it's quite pointless to discuss if looking into the envelope gives any new information, cause we don't even know the distribution yet.
If your solution is the same as this article's, it's plain wrong. Even the natural number case is plain strong.
It's very easy to demostrate as well: consider a trivia case where the distribution is just {P(1)=1/3, P(2)=1/3, P(3)=1/3} and you see 2 in the first envelope. There is no strategy to get a better chance than 50%. Therefore, any strategy that gives a better chance than 50% must implicitly make an assumption over the initial distribution (and therefore excludes a distribution like {P(1)=1/3, P(2)=1/3, P(3)=1/3})
Actually the article is even "wronger" than this, because "started A" and "switched" aren't independent and one can't simply use the product of their probability. The above example is a quick way to demonstrate it's not a general strategy without assumption to get >50% winning chance. Similarily, one can just use {P(1)=1/3, P(2)=1/3, P(3)=1/3} (this is a valid distribution over real numbers!) to demonstrate the real number strategy isn't general.
Again, for both natural number and real number case, the discussion over strategies is only meaningful is we know something about the distribution.
Interestingly, this article is wrong more or less in the same way as believing switching does give you more expected value in the original "twice money in another envelope" variation.
Edit: For people who are interested in the switching strategy, check Randomized Switching in the Two-Envelope Problem (2009). Spoiler: full of discussion over the initial distribution.
> It's very easy to demostrate as well: consider a trivia case where the distribution is just {P(1)=1/3, P(2)=1/3, P(3)=1/3} and you see 2 in the first envelope.
Seeing 2 is only one of the many possible cases. You haven’t calculated the total probability.
Sorry, but you're simply wrong. You can read the answer I gave. You can read the answer Colin Wright gave. You can trust that we both have math degrees and know what we are talking about. Or, aw heck, you can try it with an actual program at https://www.perlmonks.org/?node_id=39630. (Yes, I wrote that piece of hackery about a decade ago.)
I don't actually care how you convince yourself. But the explanation is right. If your random number is outside of the range, you've got even odds. If it is inside of the range, you've got 100% odds. As long as there is a positive probability of being between, you've got strictly better than even, by half of the probability of being between.
Many, many distributions guarantee positive odds of being in between. The one I chose for my program was:
(log(rand) * (flip_coin() ? 10 : -10 ))
Which is the log of a random number between 0 and 1, times 10 times + or - with even odds. The various factors were chosen to fit well with normal human choices that most seek to test it with.
You're correct. I realized that I completely misunderstoond the original problem after reading McDonnell's paper more carefully.
I thought you meant the strategy can make the winning chance always >50% even after the player opens the first envelope, which isn't possible.
However you actually meant the strategy can make the expected winning chance >50% before the player opens the first envelope, for any well-defined distribution of real number, even the distribution is not known to the player, which now I realize is true.
(I haven't thought through some edge case like Cantor distribution, but now I incline to it's true not just for "many distributions". Of course for a discrete distributions, we need to specifiy the two envelopes can't have the same number. Besides that, it seems to hold true for any distribution?)
Exactly right. After you've picked the envelope, you may be nearly guaranteed to be wrong. For example both numbers are large positives and you picked the smaller. Now you're almost guaranteed to guess larger, and be wrong.
But before you pick, your odds were still bigger than 50%. Just not by much.
That code has the same problem that Colin Wright's explanation does. Your computation of the success rate explicitly uses the fact that you're between the two values, but till you've seen the second value you can't possibly know whether you are or are not and thus can't adjust your guess based on that fact.
The success is checked after the decision is made to, well, verify the decision. The guess isn't adjusted on whether the number is between two values.
The strategy is straightforward and bulletproof (if you allow a random generator of real numbers, otherwise you may keep tossing coins indefinitely): keep tossing coins until you get tails. If the number you saw is less than the number of heads you got, you don't switch.
For the simplest case assume that one envelope always contains 1 and another always contains 2. You choose one envelope randomly, so in 50% of cases you get 1, which you switch in 50% of cases. And in 50% of cases you get 2, which you switch in 25% of cases. Hence, you pick the higher number in 62.5% of cases. The same works with any numbers N, M; or any complex distributions; or even real numbers with a bit more complicated strategy. You don't have to know whether you are between two values in advance, you just have to guess.
No, I'm not doing anything other than tracking it. I'm doing that to show the user, "How much of this success was chance 50% choices, how much was guaranteed?" And so people can see how much of a good or bad result was blind luck versus the strategy successfully sticking a thumb on the scales.
In other words, I'm merely trying to be informative.
You need some _known_ distribution though, and it's shocking because the distribution is ommitted from the question, and the presence of the same distribution is snuck into the answer.
Not to be too dismissive, but the title text of this xkcd [0] seems relevant. No matter how nice the explanation is, the fact that the conclusion is wrong suggests that the reasoning has a flaw. I took a stab at what I think that flaw is when I responded to your rebuttal [1].
I dimissed it at first (you can see I made some embarrassing comments above), but the strategy is correct. I'd argue it's even intuitively correct once the problem in question is clarified.
It's not saying that after the player see the number in the first envelope, the strategy guarantees a >50% outcome.
It's saying that give any distribution, over all possible outcomes, >50% times the strategy will end up pick the larger number. You can say this >50% is the expected winning chance before the player see the number in the first envelope.
I'd say this is "intuitve" because, if your strategy can guarantee "when the player see a large number in the first envelope, he's less likely to switch than if he saw a small number", it would be better than blindly switching by coin toss. So intuitively such a strategy exists.
The only "trick" here is that since the player doesn't know the initial distribution, they can't tell "how large counts as large?" therefore they needs something that preserves some property over the whole real number line. That's why the strategy involves sampling from a another distribution whose PDF is non-zero everywhere.
I was dealing with a more general variant of the version that Vos Savant gave. Which is that you were able to look at the contents of the envelope that you got. She claimed you didn't gain useful information. Calling it "information" is problematic, but it certainly is useful, because the stated strategy then works!
This is one reason that probability theorists have learned to be more careful in stating the double money version to rule out the solution that I gave. But my version has been in the literature since, I believe the 1950s. Under, as I first encountered it, the same exact name. (I encountered it from Dr Laurie Snell at Dartmouth College in the mid 1990s.)
Perhaps it'll be more intuitive to you if you scale the number of doors up. If there are 100 doors, only one containing a car, you pick one, the host reveals 98/99 remaining doors as goats, it's obvious the correct choice is to switch. The correct answer is a mathematically provable probability. 1% chance you picked the right door, 99% chance the door was in the remaining pool, therefore 99% chance the last remaining door is the correct door.
You can write a simulation that does this (as MIT and several others have) and Monte Carlo it. You will find that the logic is 100% correct. The prize MUST be either the door you picked initially or the one you can switch to. There is a 99% chance it's not the one you picked.
Can you explain the mistake (of people taking the standardized tests)? I don't understand it from the message, and the Wikipedia article seems to detail exactly the "normal, smart" switching solution.
It really isn't. The problem is usually given in this form:
> Suppose you're on a game show, and you're given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what's behind the doors, opens another door, say No. 3, which has a goat. He then says to you, "Do you want to pick door No. 2?" Is it to your advantage to switch your choice? [1]
The only correct ansewr to this question is "I don't know." People who answer yes are just taking way too many standardized tests.
[1]: https://en.wikipedia.org/wiki/Monty_Hall_problem