Jump to content

Talk:Two envelopes problem/Arguments/Archive 4

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Lowercase sigmabot III (talk | contribs) at 00:59, 4 October 2014 (Archiving 1 discussion(s) from Talk:Two envelopes problem/Arguments) (bot). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
Archive 1Archive 2Archive 3Archive 4Archive 5Archive 6

What is the paradox?

I have just thought of something that might shed some light on the disagreement on this problem.

What exactly do we think is required for the problem to be considered paradoxical? In the article itself the last steps in the argument are:

9) After the switch, I can denote that content by B and reason in exactly the same manner as above.
10) I will conclude that the most rational thing to do is to swap back again.
11) To be rational, I will thus end up swapping envelopes indefinitely.
12) As it seems more rational to open just any envelope than to swap indefinitely, we have a contradiction.

I think we all agree that for the player to swap endlessly is an absurd situation and therefore we have a clear paradox. I also suspect that everyone agrees that this paradox has been resolved for all cases.

No, this paradox has not been resolved. iNic (talk) 13:07, 25 January 2012 (UTC)
INic can you give in a new section below, or on the arguments page, an example where you think there is still an argument endless exchange. Please give a full and clear description of the exact setup and of the line of reasoning that you believe leads to a paradoxical conclusion. State when and if any person looks in their envelope. Martin Hogbin (talk) 20:10, 25 January 2012 (UTC)
Please read the Wikipedia article, it should give you the information you are looking for. iNic (talk) 17:21, 26 January 2012 (UTC)
The WP article, quite obviously. does not tell me which version you have in mind. Martin Hogbin (talk) 09:25, 31 January 2012 (UTC)
Aha in my mind? Well, in my mind there are only one paradox, only one version. All the different "versions" in the literature are due to the fact that all proposed solutions are flawed. A bad solution doesn't solve the problem if the problem is tweaked just a little bit, or slightly restated. Hence the series of never ending new "variants" of the problem. But the correct solution solves them "all" in one blow. iNic (talk) 09:36, 31 January 2012 (UTC)

At the end of his paper, Nalebuff talks of what he seems to consider a paradoxical situation that is more like the two neckties paradox. There is no suggestion that the player would swap endlessly but he does seem to consider it paradoxical that both players should find it to their advantage to swap in some cases. This seems to me to be a weaker paradox and I (and Nalebuff it seems) wonder if this paradox has been fully resolved in all cases. Martin Hogbin (talk) 11:38, 25 January 2012 (UTC)

To swap back and forth is in itself not paradoxical. Many processes go back and forth endlessly without being a paradox. Think of a pendulum for example. The paradox here is that we have rational decision theory urging us, as it seems, to do a series of rational steps that when thinking of it as a whole doesn't seem to be rational at all. This is the paradox here. How can a theory of rational behavior immediately lead to irrational behavior? The "twin construct" that is used in the neckties version is a little different. There the paradox is that something must be wrong if both persons think that they gain value by swapping their neckties. It seems to create value out of nothing. But if both gentlemen happen to think that the other tie is prettier, then they do in fact both gain if they swap. So in this case one need to be more accurate when presenting the story so that a paradox is really created. I guess this is why the common formulation of the paradox today is in terms of envelopes and money and not neckties. However, when looking in one of the the envelopes the twin construct is still useful to derive the paradox. iNic (talk) 13:07, 25 January 2012 (UTC)
In the necktie problem the gentlemen are not interested at all in which necktie is more pretty but which one cost their wives more money to buy for them. They apparently both think for the same reasons that on average they will individually get richer, in expectation value, by switching. Since they can easily imagine themselves in the position of the other person, they both think that together they will get richer on switching, yet they know that together the total cash price of their neckties did not change. So there was a paradox.

To swap back and forth imagining that each time you swap you get richer (on average) is a paradox, since you know that on switching twice you definitely have not got richer.

I think that the paradoxical nature of the different versions is pretty clear and moreover I think they have all been adequately resolved more than 20 years ago. Since then there has been a lot of repetition, a lot of re-inventing wheels, and a lot of noise. But of course ... that is just my personal opinion. On the other hand I think it is a pretty objective fact that there have not been any new "solutions" to the version (studied by mathematicians and economists) in which we imagine the writer of the TEP argument being concerned with conditional expectations since 1989, and that there have not been any new "solutions" to the version as imagined by the philosophers (and to a lesser extent psycholgoists and educationalists) who, not knowing about conditional expectation, thought that the writer was calculating an unconditional expectation value, since about 1995. There has been one claim, by Schwitzgebel and Dever, that they are the first to explain what went wrong in the philosophers' version, but I don't see any novelty in it (note: nor has their proposal been enthusiastically adopted by authoritative writers from philosophy, it has not even been discussed. Just a few papers give the citation in order to cite all known published papers in the field.)

There have been two "new" paradoxes: Smullyan's word game without probability, and Cover's trick with a randomized way to increase your expectation values when you do look in your envelope (which was already known in 1992). Smullyan's paradox has been studied in a small number of very academic papers from philosophy. Writers who are using different theories of counterfactual reasoning naturally find different solutions. But they all do find solutions.

There has been a lot of "noise" in the economics literature about infinite expectation values, in which people are endlessly arguing whether infinite expectations makes sense in real world practise. And this discussion is a repeat of the Saint Petersburg Paradox endless discussion. On the other hand this discussion says nothing about the question of where is the *logical* mistake in the sequence of apparently *logical* steps leading to a clearly nonsense conclusion. So though in the wikipedia article one can discuss this question a little, it seems to me of tangential interest. The essential paradox is resolved by demonstrating the illogical steps of the reasoning. Those steps have been well understood for a long time now, even if they happen to be different according to different interpretations of what the writer is doing. Richard Gill (talk) 14:58, 25 January 2012 (UTC)

You may find the paradoxical nature of the different versions is pretty clear but it surely would be good practice, for the benefit of laymen and experts alike, to make absolutely clear in each case exactly what is considered paradoxical. It also would seem a good idea to decide on exactly what constitutes a 'resolution' of the paradox. In my opinion a proper resolution requires identification of the error in the proposed line of reasoning not merely showing that exchange is not beneficial. I am sure that some of the noise that you complain about comes from this confusion. Martin Hogbin (talk) 10:04, 26 January 2012 (UTC)
The word "paradox" means different things in different contexts. The Wikipedia article gives a good overview. iNic (talk) 17:21, 26 January 2012 (UTC)
Yes, I know what a paradox is and I can see the paradoxical line of reasoning in WP but there are other cases where the setup and proposed paradoxical line of reasoning might be different. It never hurts to make things crystal clear and thus avoid pointless arguments. To just show that you should not swap is pointless, we all knew that at the start by reason of symmetry. Martin Hogbin (talk) 18:09, 26 January 2012 (UTC)
Exactly! This was what I tried to tell you when you used this argument yourself. I'm glad you realize now how pointless it is. iNic (talk) 16:24, 27 January 2012 (UTC)
I would be happy to point out the error in your line of reasoning of only you would give it, with the exact setup. Martin Hogbin (talk) 19:20, 27 January 2012 (UTC)
Please start right there and replace your pointless argument with a valid one. That will be interesting! iNic (talk) 09:36, 31 January 2012 (UTC)
I think the article does do just that, now, Martin. It shows for each of three or more variants of the problem, or if you prefer, different ways to imagine what the problem is, where the logical error is in the reasoning that leads to a clearly nonsensical conclusion. Now the Broome version is one which, if you are a mathematician, is easily mathematicallly (=logically) resolved, but it still remains a weird phenomenon for ordinary people. I think that is the reason why many authors slap on top of the logical analysis the further, strictly speaking superfluous, claim that in any case, infinite expectation values are stupid / unreal / whatever. Nalebuff does this, Falk does this. Many others do. Like other weird constructions in mathematics (eg Banach-Tarski) it is weird. Banach-Tarski is called a paradox because it insults our imagination and intuition, but there is no actual logical contradiction. Just an impossibility for our little brains to visualise the thing.

So I guess we can say that the Two Envelope Problem provides, among its many variants, paradoxes of both kinds: things which are logically correct but hard or impossible to imagine hence hard to believe; and things which are logically incorrect and where it suffices to point out the logical error.

I think this is a new insight in the field of TEP Studies. Thanks! Richard Gill (talk) 19:12, 26 January 2012 (UTC)

Exactly what new insight did Martin give you now? I didn't get it. iNic (talk) 10:07, 27 January 2012 (UTC)
Richard, I do not disagree with anything that you have written but you seem to have missed my point. In the article the only claimed paradox that I can see is that, by the given argument, the player should swap endlessly. Nalebuff, however sees a different paradox for some versions, namely that both players (note that there is only one in the setup given) expect to gain by swapping. These are not the same thing. Martin Hogbin (talk) 19:53, 26 January 2012 (UTC)
Yes, in the article we are copying a particular, later, statement of the problem due to Ruma Falk. It's not the original. Earlier versions tended to have two people each in possession of an envelope (wallet, box, necktie ...) and both standing to gain, on average. We, outsiders, can follow both persons' arguments so apparently must support both of their conclusions, yet we know that the total does not increase - so how come can we advise both to switch? Richard Gill (talk) 14:01, 29 January 2012 (UTC)
The new insight I got is that a problem can be solved, since we can find the error in the reasoning, but there can remain a paradox, in the sense that intuition still does not agree with the result of logical deduction. Richard Gill (talk) 14:33, 28 January 2012 (UTC)
The Broome Two Envelope Problem can be resolved, we can see where the argument breaks down. But we retain a Broome Two Envelope Paradox, since it remains counter-intuitive that if we repeatedly create two envelopes containing two numbers according to the Broome recipe, and give them to Alice and Bob, then on all those occasions that Alice has a number a in her envelope, the average of the contents of Bob's envelopes is 11a/10 > a, while simultaneously, on all those occasions that Bob has an a number b in his enveloe, the average of the contents of Alice's envelopes is 11b/10 > b; for all possible values of a and b >1. (When a or b =1, replace 11/10 by 2).

Yet the distributions of the contents of both Alice's and Bob's envelopes is the same!. Each possible value 1, 2, 4, 8, ... occurs, relatively speaking, equally often. Richard Gill (talk) 14:49, 28 January 2012 (UTC)

It's funny because Albers (2003) has exactly the opposite opinion: "The two-envelope paradox is easily explained, but the corresponding problem is more difficult." (p 29). So the total disagreement among scholars abound. Not even a simple high level description of the situation is possible to agree upon. iNic (talk) 09:55, 29 January 2012 (UTC)
These authors are colleagues of mine in the Netherlands, I know them well. When they say two-envelope "paradox" they refer to the paradoxical (in the sense of leading to a counter-intuitive result) derivation of always switching. When they talk about two-envelope "problem" they mean the problem of figuring out what the owner of Envelope A should do when he or she actually looks in his envelope first, in a real world situation. To solve this problem one needs more input. For instance: a prior and a utility and the principle of maximizing expected utility. If you accept the principle and you take a prior and a utility function as given, the problem has become soluble, it's a well defined mathematical optimization problem. But how do you figure out a prior and a utility? Anyway, why should you accept the principle of maximizing expected utility? In this sense the TE problem is not solved. You would probably agree, iNic, that in this sense it is not solved. That's my opinion too. In fact I would say it cannot be solved in general, and it's a waste of time to look for a general solution. Albers et al also say that in some situations it can be better not to give an answer at all. I too agree with this. Often the most valuable result of a statistical consultancy is that the statistician can explain to the client that his problem is ill-posed, un-answerable, and that the client should re-consider what it is he wants to do. Richard Gill (talk) 13:55, 29 January 2012 (UTC)
PS Albers et al. are not referring to the Broome paradox when they make these remarks. Richard Gill (talk) 14:03, 29 January 2012 (UTC)
OK so you and your peers in the Netherlands agree that for the Broome version the paradox is not solved but the problem is solved, while for the rest of the variants it's the other way around? Very confusing indeed. iNic (talk) 10:50, 30 January 2012 (UTC)
No. The other versions are easier. The problem is solved, there is no paradox either.

As you said yourself, the word paradox has a number of different meanings. I would like to distinguish a problem to be solved (where does an apparently logical deduction go wrong), and a situation which it counter-intuitive. Nothing to be solved, though there is something to be learnt: intuition need not be correct.

But first of all, let's define the Broome problem and paradox precisely. Suppose two numbers are generated and written down according to Broome's recipe and put in two envelopes, which are labelled completely at random (independently of the contents) A and B. The recipe is: take a coin with probability 1/3 of heads. Toss it just as long till you see the first head. Call the number of tails before the first head t. It could be 0, or 1, or 2, .... The two numbers are 2 to the power t and 2 to the power t+1. Let A and B denote the random numbers inside envelopes A and B.

Now comes the problem: what is wrong with the following reasoning? The owner of Envelope A reasons (1) given what is in his envelope, say a, the expected contents of the other envelope are 2a if a=1 and otherwise 11a/10; in any case, it is strictly larger than a. He reasons that therefore, (2), without looking in his envelope, he will become better off by switching to the other envelope.

The conclusion is obviously false since the marginal probability distributions of A and B are identical. Exchanging one closed envelope to the other does not improve his situation at all - it leaves his situation identically the same to what it first was, since the bivariate probability distribution of the ordered pair A,B is the same as the bivariate probability distribution of the ordered pair B,A.

Our problem, as I said, is to explain where the apparently logical reasoning breaks down. The answer is that it breaks down at step (2). In step (2) we are silently using a theorem which requires a condition which is not satisfied in this case. The conditions of the theorem do not apply hence we cannot deduce that the conclusion of the theorem is true.

Still, we are left with the paradoxical (in the sense of counter intuitive) fact that in many, many repetitions, for any particular amount in Envelope A, the average of what's in Envelope B is larger, and vice versa.

Problem solved, paradox remains. The paradox that remains does not have a solution and does not need a solution. We have learnt that our intuition about averages can break down when infinities are involved. This has been known for several hundred years. This is not a paradox which can be solved, but it is a situation to which one can become accustomed ... in other words, one can develop a better intuition.

Von Neumann said: one does not come to understand new mathematics, one gets used to it. This is the sort of paradox which goes away by being aware of the possibility and getting used to it. It doesn't go away by logical deduction.

Unfortunately, many writers on TEP are not happy with this situation. They use overkill by trying to argue that the Broome situation could never turn up in the real world. In other words: sorry guys, we are stuck with the paradox, but don't worry your little heads about it, you won't ever have to deal with it in practice. Richard Gill (talk) 13:43, 31 January 2012 (UTC)

Richard I agree with what you say and I have a suggestion that might help in your mathematicians vs the rest argument. When talking about the finite, physically reasonable cases it is alright to use a real physical representation of the problem, in particular to use terminology like 'envelope, and 'money'. When referring to the unbounded versions I think it would be better to use strictly mathematical terminology, 'set', 'number' etc.
It is quite right to say that no one will never have to deal with the untruncated Broome game in practice. Infinite amounts of money do not exist, neither do envelopes large enough to contain them. On the other hand, we can deal with infinities mathematically.
The problem is caused by a desire to glamourise the more spectacular results of mathematics. In the formal section on the Banach-Tarski paradox [[1]] we see the problem expressed mathematically. The problem is that it does not look very exciting so we are tempted to talk of gold balls which can be doubled in size. Of course, real gold balls cannot be doubled but the mathematics is perfectly correct. The same problem exists in QM where writers attempt to explain QM results in terms of classical physics. This produces exciting paradoxes but is not particularly helpful in understanding the subject. Martin Hogbin (talk) 18:19, 31 January 2012 (UTC)
Richard, you say that "In step (2) we are silently using a theorem which requires a condition which is not satisfied in this case. The conditions of the theorem do not apply hence we cannot deduce that the conclusion of the theorem is true." What theorem is that? iNic (talk) 10:23, 1 February 2012 (UTC)
Theorem. Suppose E(X) is finite. If E(Y | X=x) > x for all x, then E(Y) > E(X). Richard Gill (talk) 20:58, 7 February 2012 (UTC)

The 'simple solutions' explained

The simple solutions, as exemplified by the solution given by Falk in her 2008 paper, are correct but only in particular circumstances which have not thus far been clearly identified. In her paper, Falk says:

The assertion in no. 6 (based on the notation of no. 1) can be paraphrased ‘whatever the amount in my envelope, the other envelope contains twice that amount with probability 1/2 and half that amount with probability 1/2’. The expected value of the money in the other envelope (no. 7) results from this statement and leads to the paradox. The fault is in the word whatever, which is equivalent to ‘for every A’.

So far this resolution is in accordance with Nalebuff's solution.

She continues:

This is wrong because the other envelope contains twice my amount only if mine is the smaller amount; conversely, it contains half my amount only if mine is the larger amount. Hence each of the two terms in the formula in no. 7 applies to another value, yet both are denoted by A. In Rawling’s (1994) words, in doing so one commits the ‘cardinal sin of algebraic equivocation’ (p. 100).

It is not quite clear to me exactly what this bit is saying. What exactly is, 'the smaller amount'?

Falk explains further by giving a revised version of the paradoxical line of reasoning:

1. I denote by A the amount in my selected envelope.

2′. Denote by S the smaller of the two amounts.

The probability that A = S is 1/2 and that A = 2S is also 1/2.

3′. The other envelope may contain either 2S or S.

4′. If A=S the other envelope contains 2S.

5′. If A=2S the other envelope contains S.

6′. Thus the other envelope contains 2S with probability 1/2 and S with probability 1/2.

7′. So the expected value of the money in the other envelope is 1/2 2S+1/2S=3/2S.

8′. This is equal to the expected value of A, which is 1/2 S+1/2 2S=3/2S, so I gain nothing on average by swapping, and I am indifferent between keeping and swapping.

My assertion is that this argument, and many similar ones, are perfectly correct iff 'the smaller amount', as represented by S above, is a constant. In other words it has a fixed value. This means essentially that the solver of the problem knows the possible sums in both envelopes. In other words, our envelope space contains only two elements.

Later Falk says:

Had step 2 been adjusted to say ‘the probability that A is the smaller amount, denoted As, is 1/2, and that it is the larger amount, denoted Al , is also 1/2’, then the same symbol A would not have been used for the two cases in steps 6 and 7, and the paradox would have been avoided.

What are As and Al here, random variables? If so the argument above does not really make sense. On the other hand if they are both constants everything is clear.

In the Broome version of the game, because of the way the sums in the envelopes are determined, it is not true that the sum in the unchosen envelope is equally likely to be half or twice the sum in the chosen one - unless, after the two sums in the envelopes have been determined, we are told what they are, reducing our envelope space to only two elements.

In my opinion this is the crucial fact missed by many of the authors of papers on the subject. You cannot ignore how the sums in the envelopes are determined in your calculations unless, after they have been determined, you are told both sums. What you can never, in my opinion do, is have variable constants.

Richard, I would be interested to hear your view on this observation. Prof Falk, if you are reading this, I am sorry for using your paper as an example of what I consider to be incomplete thinking, but I would be interested to hear from you also.Martin Hogbin (talk) 10:22, 8 February 2012 (UTC)

I agree. Except, that the analysis does not require the amounts to be known. We can do probability calculations conditional on the values of the two sums of money (the smaller and the larger), without knowing them. I am afraid I have had so many fruitless discussions with amateurs about the fact that one can usefully think about E(Y|X=x), the expectation value of one random variable given that another takes on a particular value, without that value actually being known or given. Some of those amateurs were professional physicists, who certainly could do mathematics better than me, in the sense of having a fantastic intuition for finding really smart ways to do extremely difficult calculations. Richard Gill (talk) 10:51, 9 February 2012 (UTC)
Some more thoughts on this. The word "variable" has a very precise meaning within rigorous (precise, formal, academic, professional, theoretical, abstract, ... ) mathematics. And then always accompanied by a quantifier "for all" or "there exists". Examples: for all x in some set, x satisfies some relation; or: there exists x in some set, such that .... Brackets indicate the precise scope of the quantifier. The word "constant" is not used, and a word like "fixed" or "known" also has no meaning. Careful mathematics is like careful computer programming. Mathematical variables are just like variables in computer code, they are merely place holders, and one is careful to show which places are supposed to be occupied by the same thing (scope; local versus global variables), and what this thing can be (type).

Within probability theory, but only within probability theory (and branches of mathematics built on probability theory, like statistics) we use the words "random variable" and "constant" with rigorous, precise, formal meanings. Within probability theory the notion of conditioning is useful. One can say things like "the probability of A, given B" but this does not mean that B is known to have occured. It simply means (if you like frequentist language) "imagine many, many repetitions of the probability experiment we are talking about. Restrict attention to all those occasions, and only those occasions, when B occurs (completely irrelevant whether anyone knows this to the case, or not). What is the relative frequency of A within this sub-family of repetitions?".

Unfortunately, ordinary folk think that variables must vary and constants must be fixed and possibly also known, and that we can only talk about the probability of A given B if B is known (known to someone, to have happened). But in probability theory, a constant is merely a random variable with a degenerate probability distribution. A random variable is a (deterministic) function from the set of possible atomic outcomes of some probability experiment to some set of possible values, typically real numbers. Conditioning on an event, does not mean that we force the event to happen. Working given X=x does not mean that we know or force X to take the value x. It merely means a conceptual restriction to all those events, and only those events, when the random variable X happens to take the value x. The symbol x denotes a mathematical variable. The context must show what it can be, and whether the name x is only being used locally or whether it is being used globally (or something in between).

Another word with precise technical meanings, various ones in fact, within mathematics, is "independent". Another recipe for confusion. Richard Gill (talk) 09:28, 10 February 2012 (UTC)

The philosopher's stone

[Moved from the main discussion page.]

I think you have missed a point though. You say, 'The solution above doesn't explain what's wrong if the player is allowed to open the first envelope before being offered the option to switch', but it also does not explain what is wrong if the first envelope is chosen from a non-uniform distribution, such as in the Broome game. We both know that the probability that the original envelope will hold the smaller sum is not, in that case 1/2. The 'common resolution' does work if there are only two possible sums in the envelopes (or if after they are chosen we are told what they are) in other cases it is nonsense. Martin Hogbin (talk) 19:16, 8 February 2012 (UTC)
The way I think about this (i.e., according to my own POV), the common resolution is not a resolution at all. But the second resolution is completely adequate. And the way I think about it, the Broome paradox is a completley new paradox, built on the second resolution. You thought you had flattened the bump in the carpet but then someone changes the rules a little, and a new bump came up behind your back. Broome does not suddenly make all the steps of the original TEP argument correct. We have just argued why they can't all be correct! Broome requires us first to rewrite the steps, with a correct computation, for his rather special circumstances, of the conditional expectation of what's in B given what's in A (which involves probabilities 3/5 and 2/5, not 1/2 and 1/2).

I think that Broome can justly be called a new variant. It is a different paradox, but clearly inspired by the original. The original is the original... (Nalebuff, Gardner). The original has been interpreted in two different ways in the literature: the second interpretation was the original interpretation, the first interpretation became dominant in the philosophy literature and to some extent the popular literature.

To make things more complicated still, Nalebuff already had a Broome-type example, in order to show that maybe the solution of our second interpretation isn't the whole story.

All in all, two envelopes problem is actually a rich family of related paradoxes; some are "just" logical paradoxes, easy to resolve by showing the incorrect logical step; others are brain-twisters, things which are hard to imagine but which can actually be true. These things which force us to expand our minds do have impact on thinking in applied fields, because the challenge to intuition needs to be addressed from the practical point of view. Richard Gill (talk) 20:07, 8 February 2012 (UTC)

I think you have missed my point. The common resolution makes sense in so far as it shows that the unconditional expectation in both envelopes is the same but it only does this if there are only two possible sums in the envelopes.
If the first envelope is chosen from, say, the set {2,4,6} the common resolution does not even calculate the unconditional expectation. Do you see what I mean? Martin Hogbin (talk) 22:24, 8 February 2012 (UTC)
I think I see what you mean, and I think in the common resolution the author is silently taking the two amounts of money in the two envelopes to be fixed. The writer does not distinguish random variables from constants. He writes A and X but he should write A and x. It takes years to teach students to distinguish between capital letters and small letters. Nowadays handwriting is so bad they don't even make any distinction in their ordinary writing.

There would be no TEP if the writer had carefully made the distinction, as many reliable sources have said!

So I have been bold again. I have added an expanded and mathematically respectable version of this common resolution as a sub-sub section to make all this explicit. Similarly I added more mathematical details of the second resolution as a sub-sub section. I hope in this way both laypersons can be satisfied, and mathematicians can be satisfied. The two resolutions are now written out as parallel stories which only differ (probabilistically speaking) according to whether we are conditioning on the pair of sums of money (e.g., on the value of the smaller amount), or we are conditioning on the amount in Envelope A.

For a probabilist this is the important distinction. The layperson does not really know this distinction. Instead, the layperson thinks of things being known and things being not known. Unfortunately, this puts the lay person on the wrong track.

The layperson who has a nodding acquaintance with elementary algebra and calculus thinks that there is an important distinction between variables and constants. True, the distinction is important, but here, another distinction is important. The distinction between random variables and possible values of random variables. Whether those possible values of random variables are thought of as constants or variables, and whether they are actually known or not known, is almost irrelevant.

I think here we are at the heart of the problems of the bifurcation of TEP into a philosopher's problem and a probabilists's problem. Many people have quite a lot of mathematical training but at the same time almost no training in probability. Such people can easily see TEP in the philosopher's way rather than the probabilist's way. An interesting example is given by Ruma Falk, who has taken both points of view in different papers over the years, with different co-authors. By training she is an educationalist and a psychologist, and not a statistician or a probabilist. Though she has written extensively on probability problems it seems to me that she writes more as a mathematician than as a probabilist, so occasionally making the wrong distinctions, and not using the optimal notation. Richard Gill (talk) 10:37, 9 February 2012 (UTC)

Richard, is the explanation of how the simple solutions arise and the precise problem that they do answer anywhere in the literature? It may seem obvious now but I have not seen this explanation in the literature that I have looked at. Martin Hogbin (talk) 16:33, 9 February 2012 (UTC)
Very good question. I am hoping that the paper by Thomas Bruss, which I have not yet read, does this. He is a well known fundamental mathematics oriented mathematical statistician, and statistical decision theorist. So I can't imagine that he has written something superficial or even wrong on TEP. I mailed him, and he has posted me an offprint of his 2006 paper. At the moment, I admit, this is an original synthesis by me (though it is completely elementary).

On the other hand, you could say that I am merely expressing in modern probability formalism, what a number of philosophers have said in lengthy articles full of long words and difficult philosophical concepts. I merely abstracted the mathematical essentials from their words. Made sense of their verbal stories in the only way possible which makes mathematical sense, hence presumably is indeed the intended meaning. These guys write papers which are bloody hard to read, but on the whole they do not actually write nonsense,Richard Gill (talk) 19:09, 9 February 2012 (UTC)

That is not quite what I was getting at. I have not seen this explanation offered anywhere until I mentioned it just above. Martin Hogbin (talk) 19:57, 9 February 2012 (UTC)
I think you have seen these explanations in my own draft paper! But writing for mathematicians you possibly did not catch what I was saying. By the way, Rawling is definitely thinking of the smaller amount as being fixed, he calls it s. And by his analysis the probabilities 1/2, 1/2, are correct, it is the substituted values which are wrong. So his analysis, with high probability, corresponds to what we now have.

He also uses words like "syncategorematic". I suspect that none of us actually understands his paper at all, but just grabs at a few fragments which seem recognisable. I suspect that many writers who cite him also hardly understand a word of what he says, except for the phrase "the cardinal sin of equivocation", which people love to quote. I suspect that none of them even saw the word "equivocation" before. (I think this particular paper really is a load of BS). Richard Gill (talk) 20:30, 9 February 2012 (UTC)

I think we must be talking about slightly different things. In your paper, you say, According to [the philosophers'] interpretation we are aiming at computation of E(B)... but that is just crazy. However we may propose to arrive at it a computation of E(B) is completely pointless. As I think you have pointed out elsewhere, it is a fixed number, there is absolutely no point in comparing this with random variable. I cannot believe that the philosophers are being this stupid. What I am saying that they do is that they are conditioning everything, in your notation, on X=x (or Y=y) right at the start of the problem. Martin Hogbin (talk) 09:47, 10 February 2012 (UTC)


Probably we are talking about different things. The philosophers are not mathematicians. My point of view is that the simple resolution does not assume the amounts in the envelopes to be known and fixed. It merely assumes that we are doing a probability calculation conditional (in the probabilistic sense) on X=x, where X denotes the smaller of the two sums, and x is a possible value thereof. The correct calculation shows that E(B|X=x)=3x/2. This statement is conventionally abbreviated to E(B|X)=3X/2. By averaging again, over all possible values x of X, one recovers E(B)=3E(X)/2. As a special case, X might be imagined to be constant. As an even more special case, it might be supposed to be known. Doesn't make any difference. Also, it doesn't make a difference whether probability is taken in subjective or frequentist sense, or as a mixture of the two. A frequentist might typically take X to be a constant random variable, equal to the fixed value x. Denoting the smaller amount by x, the name of a mathematical variable, does not imply that we think of x as being known or unknown, constant or varying. We simply to a calculation which applies to any situation, whatever x might be. Richard Gill (talk) 09:53, 10 February 2012 (UTC)
PS what amazes me is that the philosophers do completely fail to notice that there is another big thing wrong in their calculation. They are obviously computing an expectation (maybe a conditional expectation). But they end up on the right hand side of their equation with A, they seem to have forgotten that according to an earlier step A was apparently a random variable. So maybe quite a few of them are exactly that "stupid" as not to notice this. Richard Gill (talk) 09:56, 10 February 2012 (UTC)
We probably should be having this discussion somewhere else, on the arguments page for example.
I do not quite understand what you are saying. You say above, ' My point of view is that the simple resolution does not assume the amounts in the envelopes to be known and fixed', but you go onto say, 'It merely assumes that we are doing a probability calculation conditional (in the probabilistic sense) on X=x, where X denotes the smaller of the two sums, and x is a possible value thereof'. Is that not the same thing? If we condition on X=x then the smaller sum (and hence the larger sum) has a known fixed value. Martin Hogbin (talk) 10:14, 10 February 2012 (UTC)
One can meaningfully compute probabilities and expectations conditional on X=x, with or without actually knowing x. In fact that is the whole point of the second interpretation. Without looking in Envelope A we can imagine looking and seeing A=a. The symbol a is a mathematical variable. It stands for some number, some amount of money. Any amount of money. Similarly, doing probability calculations conditional on X=x carries no implication whatsoever that someone in this story actually knows x. That's the whole point of use of mathematical variables in doing algebra. You can say "let x be the solution" without knowing x and even without knowing if there is a solution or not, or many solutions or only one. Richard Gill (talk) 11:36, 10 February 2012 (UTC)

OK then I think I means something different from you. Let me explain exactly what i mean in two different ways then you can tell me how to express this more formally.

From a frequentist perspective we imagine the experiment to be repeated many times. The sum in one envelope is determined by some process (for example uniformly at random from a fixed sample space or by the Broome method) then double that sum is put in another envelope. Then a fair coin is tossed and according to the fall of the coin the player receives one of the envelopes.

If that process is repeated from the coin-toss stage only, that is to say with the same starting envelopes then, in my opinion, the philosophers' solutions are fine. From a Bayesian perspective this equates to the player being told the sums in the two envelopes, even though they may change from game to game.

This is not quite the same as saying that the sums in the envelopes are fixed, they are always fixed; nobody changes their value at any time. However, if they are fixed and known then Falk's resolution, for example, makes sense. Martin Hogbin (talk) 13:13, 10 February 2012 (UTC)

Yes, I agree! Falk's solution silently assumes we are thinking (in frequentist terms) of repetitions in which only the labeling (A and B) of the envelopes varies from repetition to repetition. For a probabilist this is the same as saying either that X and Y are taken to be constants (as opposed to random variables) from the outset (so maybe better to use lower case x, y); or that we are probabilistically conditioning on particular values X=x, Y=y. There's no mathematical difference, from here on. (Mathematically, the second possibility is more general - it includes the first).

The second resolution, thinking of A being fixed, you could say, makes more sense when X and Y are at the outset thought of as random (e.g., with probability distributions representing our prior uncertainty as to their actual values). But it seems that for many philosophers this was too sophisticated to contemplate. And possibly for many amateurs, too. That would explain the popularity of a resolution like Falk's. She can explain it much better to a class of young educational psychologists than the second. Richard Gill (talk) 10:50, 11 February 2012 (UTC)

I thought this might be an important point to make in the article but I am no so sure now. See below.

Why we will never find the philosopher's stone

My turn now to agree with you Richard. The literature (and consequently this article) is a mess. I think the reason is clear and returns us to one of the fundamental issues of statistics and probability, that without a well-defined question there can be no clear answer.

We start with an ill-posed setup. We then have a bogus proposal which we first have to interpret and then find the flaw in our own interpretation. Nalbuff (and maybe a few others) rounds up all this chaos and, giving reasonable interpretations of what it is supposed to mean, pretty well resolves it all.

Other writers, often philosophers, then come along and give alternative resolutions but invariably without saying exactly what paradox they are trying to resolve. Many produce long and mathematically dubious arguments to show something we knew from the start, that we should not swap. The aim should be not to persuade ourselves that endless swapping is pointless, that is obvious, the objective is to find the flaw in a proposed and plausible argument that we should. This, to my mind, requires that we should first clearly state the bogus argument and then show exactly where the error in that argument lies. Most of the literature completely fails to do this and works on the basis that everybody knows exactly what the paradox is. That is false; if it were true there would be no paradox.

So what to do with the article? It is a mess because Richard has made it reflect the literature on the subject. That is fine but it is of little value to the general reader. I would like to add a section for the general reader, at the start of the solutions section, based mainly on Nalebuff's paper on the subject and giving simple, plain language but mathematically sound resolutions for the paradox given at the start of the article. I would hope to get a consensus for this. Martin Hogbin (talk) 12:25, 11 February 2012 (UTC)

Personally I think the article is in quite good shape now. But then, I'm someone who (by now) already knows the literature and moreover has specialist interest in the technical niceties. So maybe it's a good encyclopedia article for an encyclopedia of statistics, but not for a general encyclopedia.

Maybe the comprehensive survey of the literature should be preceded by a short layman's article, restricting attention to highlights only, and concentrating on the original TEP (what we now call the second interpretation).

Why don't you just be bold and draft such an article, Martin? It could also go in Simple Wikipedia. Richard Gill (talk) 15:17, 11 February 2012 (UTC)

Sorry Richard, I meant it was incomprehensible for the layman. As an expert discussion I am sure it is fine. Martin Hogbin (talk) 00:49, 12 February 2012 (UTC)
No apologies needed. I agree with you! Be bold... Richard Gill (talk) 09:10, 12 February 2012 (UTC)
I like the new Intro !!!! Richard Gill (talk) 18:06, 13 February 2012 (UTC)
Thanks. I think it provides an explanation for the general reader. Martin Hogbin (talk) 23:02, 13 February 2012 (UTC)

This intro is bad in so many ways that I don't know where to begin. Best thing to do would be to delete it and improve the article instead. Make it readable and fun again, as it was several years ago. iNic (talk) 01:33, 14 February 2012 (UTC)

Perhaps you could start by saying what you think is factually incorrect about it. It is what the sources say. Martin Hogbin (talk) 09:40, 14 February 2012 (UTC)

It was placed in the same section as the "A common resolution," yet that solution wasn't even mentioned in your intro. Weird. The non probabilistic version wasn't mentioned either, of course. It's totally fine that you and Richard doesn't like or understand these ideas personally but it's a clear violation of NPOV to leave them out as you two constantly do (but Richard has improved in this respect lately). In addition you introduce a lot of Bayesian concepts in your intro without any explanation whatsoever what it is or why some authors think these concepts are needed and useful in this context. In particular they are not needed at all in the first "A common resolution" that appeared directly after your intro. They are needed in other common interpretations of the problem, but not all. This need to be explained, not just taken out of the blue. If we make the entire page more readable an intro of this kind is not needed. If written in this way it confuses more than what it helps. iNic (talk) 10:13, 14 February 2012 (UTC)

As I am sure Richard will agree, the 'common resolution' that we have now is not really a resolution of the stated paradox at all, it does not say where the error is in the proposed line of reasoning, it does not even use the same notation. It is just a reason why you should not switch, which is obvious.
My introduction is not a summary of all the resolutions but a statement of the most common resolutions from mathematicians. If you want a summary of the whole subject, the place for this is the lead. That could do with rewriting anyway. Martin Hogbin (talk) 17:04, 14 February 2012 (UTC)
So why don't you say that then? Replace the header "Introduction" with "My personal summary of all the resolutions I think are the most common from mathematicians." Then please sign your section so that all new readers will know who's view it is. iNic (talk) 10:13, 15 February 2012 (UTC)
This afternoon I was discussing TEP with two mathematicians with quite strong interests in philosophy. One of them said: the difference between mathematicians and philosophers is the following. All a mathematician needs is lots of pens and paper and a rubbish bin. All a philosopher needs is lots of pens and paper.

I plan to give TEP (wikipedia) a break for a while. It will always be hard to write a satisfactory article since the literature is so large, so much of it is very low quality, and there always will be people who look at is from a more mathematical point of view and people who look at it from a more philosophical point of view, and honestly, I think it may take another hundred years before these groups start to understand one another's point of view. There do not exist previously existing surveys of the whole literature which are both authoritative and neutral. This one on wikipedia is the first, ever! Let some new editors get involved. Richard Gill (talk) 21:08, 14 February 2012 (UTC)

Yes I agree. More editors need to be involved. Both with and without a rubbish bin. (At last I understand why it was such a big quarrel about the right to have a rubbish bin between Apple and MS a long time ago. Thank you!) iNic (talk) 10:13, 15 February 2012 (UTC)
PS we were talking about decision theory. The standard von Neumann-Morgenstern theory based on utility functions is simply no use for the real world. Suppose you were offered either 2 M Euro, or a 50/50 chance of getting either 0 or 5 M Euro. Unless you are Bill Gates, you would probably prefer the former. Suppose on the other hand you were offered either a lottery ticket which has a one in a million chance of winning two million Euros, or a lottery ticket which has a one in two million chance of winning five million Euros. Almost everyone would prefer the latter. Its a mathematical theorem that there does not exist a utility function (expressing the value of money) such that both of these preferences follow by supposing that people want to maximize expected utility. This is the Allais paradox. This is why "usual" Bayesian decision theory is a waste of time. Not only are the preferences I just mentioned the ones which most people actually do make in practice, I think we would mostly agree that they are entirely rational, too. Richard Gill (talk) 21:17, 14 February 2012 (UTC)
Yes Allais showed that the standard von Neumann-Morgenstern theory is flawed. But Allais didn't show exactly in what way it is flawed, how the theory can be fixed. The defenders of the theory has said that even if we have anomalies like this the vN-M theory is anyway the best decision theory we ever can get because any deviation from their axioms would lead to even larger irrational behavior. The interesting thing about TEP is that it shows the way to a real solution to this old problem, which Allais' problem never did. iNic (talk) 10:13, 15 February 2012 (UTC)
Well, according to the experts I talked to yesterday, there are well known ways to generalize vN-M so that Allais is no longer an anomaly. For instance John Quiggin's Rank-dependent expected utility. I think they correspond well to human psychology too. Of course the problem you get with generalizing a theory is that it fits everything and explains nothing. There are so many free parameters that you can always make it "explain" what you have seen in the past. The question is, is it useful for predicting things in the future? I don't know. Richard Gill (talk) 10:59, 15 February 2012 (UTC)
Yes true, but in these cases the resulting theory is purely descriptive. No trace of the rationality ingredient—that was so important for the early probabilists—are left. The vN-M theory is the last attempt to create a theory for the rational man. But I'm not an expert in this field so there might be later theories that I'm unaware of. I for sure need to read more in this area. iNic (talk) 14:38, 15 February 2012 (UTC)
As I understand it, Quiggins' approach has two components: utility is how one values money, but there is a different and new component which describes how one values risk. Nothing irrational in enjoying small chances of winning large amounts, and at the same time "playing safe" concerning large amounts. If only the bankers had behaved that way with our money! So one does not abandon rationality. On the contrary, the vN-M characterization of rationality simply did not do the job. I think Keynes already emphasized this. And vN-M were not dogmatic: they started with something simple. Before there was nothing. We now know their approach is inadequate. But still it forms a platform for further developments. Richard Gill (talk) 07:06, 17 February 2012 (UTC)
OK but for every new "refined" theory the concept of rationality becomes more and more diluted. It has now reached homeopathic levels of dilution. We can't distinguish a rational behavior from an irrational anymore using these theories. Everything goes. Any behavior can be "rationalized" afterwards as the theory has so many free parameters. I would say that not a single atom of the original ingredient of rationality is still left in this new homeopathic concept. Try to use Quiggins' theory during an evening at your local casino and you'll understand what I mean. iNic (talk) 14:13, 17 February 2012 (UTC) 
Sure, you don't have to tell me. Richard Gill (talk) 11:19, 18 February 2012 (UTC)

This is not a paradox.

Let us say you are given two indistinguishable envelopes, each of which contains a positive sum of money. One envelope contains twice as much as the other. You may pick one envelope and keep whatever amount it contains. You pick one envelope at random but before you open it you are offered the possibility to take the other envelope instead. You therefore swap envelopes.

The above statement is not a paradox. Nobody sees a reason to swap until a bogus line of reasoning is presented to them which appears to show a rationale for swapping. Consequently, just showing that you should not swap cannot be a resolution of a paradox. For a paradox to be resolved there must be one in the first place.

To resolve the TEP you must first state a paradoxical line of reasoning then show what is wrong with it. Anything else is pointless waffle. Martin Hogbin (talk) 09:40, 29 February 2012 (UTC)

Exactly, Martin. To resolve the TEP you must first state a paradoxical line of reasoning, and then show what is wrong with it. Literature in abundance.
If you take the two "colours" of red and green, then B can be red only under the condition that A is green, and B only can be green under the condition that A is red. Either, or, and no way out. Without any reference whatsoever to "values" or to "amounts".
In the switching argument of the TEP lines 1 to 6 are fully correct only for the asymmetric case that A clearly represents the determining amount, and B clearly just represents the derived, the dependent amount. That B entirely and fully depends on a predetermined A.
And lines 1 to 6 may also be "valid" for the case that there should be symmetry, but only as long as no one considers those lines to imply any asymmetry of any "determining" and "dependent". So these lines are very incomplete and too imprecisely worded to help prevent fallacies. For in the symmetric case, without any "determining" and "dependent" amounts,
the other envelope B can never contain 2A unless A = B/2, and
the other envelope B can never contain A/2 unless A = 2B. Just an excluding "Either - Or".
Meaning that "B = 2A" only can be possible if "A = B/2" (as it is with red and green), and "B = A/2" only can be possible if A = 2B. Just an excluding "Either - Or", and never any "as well as", like in the asymmetric case, adding both and dividing by 2.
So it is a complete illusion to conclude that B ever could be 5/4 A in the symmetric case.
To clearly see that, take my example of "one envelope contains the 200 fold amount of the other envelope." Then, supposed A determines B, switching from A to B will give you on average the 100.0025 fold amount of A, and switching from B to A will give you B/100.0025. But in the symmetric case, like "red and green", all you ever will know is that both, A and B likewise, can be expected to be (A+B)/2. And nothing else.
So at least line 7 is "the flaw", in misinterpreting lines 1 to 6 to mean the asymmetric case only of a strictly "determining A" and some strictly "dependent B". Literature in abundance. Gerhardvalentin (talk) 15:52, 8 March 2012 (UTC)

Please have a look to my attempt to distinguish "either/or" from "as well as". Gerhardvalentin (talk) 08:30, 30 March 2012 (UTC)

A possible fault

The 'flaw' I see is that by line 9 (switch, assume content to be B), people discard the previous lines 1-8. Line 9, following the same train of thought of lines 1-8, should give rise to the following:

9.1) Denote by B the amount in the currently selected envelope.
9.2) The probability that B is the smaller amount is 1/2, and that it is the larger amount is also 1/2.
9.3) If B is the smaller amount, then A is the larger amount: (A,B=A/2).
9.4) If B is the larger amount, then A is the smaller amount: (A,B=2A).
9.5) Thus the value in the previously selected envelope is A. This is smaller than the expected value of B (A*5/4).
10) Therefore, conclude that there is no further need to swap.

I think the paradox works that way: By discarding lines 1-8, one starts the entire argument from the beginning. By discarding lines 1-8, the important relationship between A and B is not established (that B is only either 2A or A/2). There may be other errors in the argument, but this is what I see as the central flaw of the argument. 112.205.100.28 (talk) 13:51, 24 July 2012 (UTC)

The stated problem ignores the value of the chosen envelope, which is lost on swapping.

You have selected one of two envelopes, it contains X or 2X each with a probability of ½. If you swap to the other envelope then you lose the contents of the envelope you are holding and gain the contents of the other one. So with a probability of ½ you lose X and gain 2X or with a probability of ½ you lose 2X and gain X. Your expected gain or loss is:

½(-X + 2X) + ½(-2X + X) = 0

In the long run, there is no gain or loss by swapping. --Norman Sheldon (talk) 08:31, 3 September 2012 (UTC)

I think this is the correct solution to the so-call paradox, because when you switch you expects gains not value of your envelope which is given to you at the first place. - Tom Cheung

Random Question

Wouldn't something along the lines of there being a cost (in terms of time) to switching also fix this problem? Regardless of how much money is in the envelope, wouldn't you eventually want to be doing something with your life than sitting there switching envelopes? Subtracting some amount, c (cost), with every switch should eventually make their utility from the problem negative. Jabberwockgee (talk) 19:59, 21 May 2012 (UTC)

Critique 1

I will switch if I selected the envelope with the smaller amount or stick if I selected the envelope with the larger amount. Knowing an amount only helps if it provides a bias between which one I selected (envelope with the smaller amount or the envelope with the larger amount).

Consider an adjustment to the basic set up that gives you even more information. You are told that there is a population of many pairs of envelopes such that one envelope contains twice as much as the other and your pair has been randomly drawn from this population. Together with the pair of indistinguishable envelopes you are given a spread sheet. The spread sheet lists every envelope pair in the population from which your pair has been randomly drawn, shows the number of each of these pairs in the population and (so you do not have to calculate it yourself) the percentage frequency this pair has within the whole population.

1. You have all the distribution data that you need for amounts in pairs of envelops, do you switch envelopes?

2. You open your selected envelope. It contains A; you are invited to switch. Now if the value A only appears in one place in the spread sheet, then you can determine if it is in the envelope with the smaller amount or the envelope with the larger amount. Now it is clear whether to switch or stick. Otherwise it appears twice, once as the smaller amount (in the pair A and 2A) and once as the larger amount (in the pair A/2 and A). Switching in the first gives a gain but switching in the second gives a loss. The percentage frequencies in the spread sheet will help you decide if you expect to gain or lose by switching. There is no guarantee that you made the right decision in this arbitrage.

3. You are told that one of the two envelopes contains 20 dollars, do you switch?

4. You are told that one envelope contains 20 dollars and the other envelope contains 40 dollars, do you switch?

Instead of the basic set up, you are given two indistinguishable envelopes. You are told that one envelope contains a US Government Bearer Bond to the value of 100,000,000 dollars and the other contains several sheets of used lavatory paper. Do you switch? Well the question you have to ask yourself is “Do I feel lucky?” Well, do ya, punk?

--Norman Sheldon (talk) 10:39, 12 September 2012 (UTC)

Critique 2

Specifically, in the switching argument, if I accept steps 1 to 8 as valid then step 9 is invalid. After the switch I agree that I can denote the contents by B, however I have arrived at this step by nominating the original envelope as containing A. The only values B can take that are compatible with this are 2A or A/2. Each has a probability of ½ so my expected value for the holding is 1.25A which is greater than A and I will not switch again.

However, I cannot accept steps 1 to 8 as valid. This is because the steps do not identify the correct entities to evaluate. The basic set up starts with two indistinguishable envelopes, i.e. an envelope pair drawn from a population of envelope pairs, yet all further analysis concerns itself with an individual envelope but not the pair as an entity. An envelope pair is made up of two envelopes whose contents are positive amounts such that the smaller amount is X and the larger amount is 2X. X is a variable whose value can be any of the smaller amounts available from an envelope pair in the population of envelope pairs. The probability distribution for X (i.e. the smaller amount) is given by P(X), hence the probability of the pair 2A and A is given by P(X=A) and the probability of the pair A and A/2 is given by P(X=A/2).

The switching argument appears to be an attempt to decide to switch when the selected envelope’s content is revealed and then the switch is offered. It goes off track from step 1.

Step 1 denotes the amount in my selected envelope as A. Remember that we have an envelope pair, the envelope holding the smaller amount contains X and the envelope holding the larger amount contains 2X. If we are solving the given question, then if A is in the envelope containing the smaller amount then A=X (and the other envelope contains 2X); if A is in the envelope containing the larger amount then A=2X (and the other envelope contains X). So let us see how the argument develops.

Step 2 says “The probability that A is the smaller amount is 1/2, and that it is the larger amount is also 1/2.” More correctly for the given question, this should say the probability that my selected envelope contains the smaller amount is 1/2 and that it contains the larger amount is also 1/2 i.e. the probability that A=X is 1/2 and the probability that A=2X is also 1/2.

Step 3 says “The other envelope may contain either 2A or A/2”. This is step of logic too far for the given question. This step is saying that we have an envelope pair containing A and 2A or an envelope pair containing A/2 and A i.e. we are considering two separate envelope pairs. This is a strong hint that argument is considering switching after revealing the content.

Steps 4 and 5 add nothing to the argument.

Step 6 is the point to stop and go through this.

I am given an envelope pair containing two envelopes etc. I assign a value A to the content of my randomly selected envelope. I have selected the envelope containing the smaller amount with a probability of 1/2 or the envelope containing the larger amount with a probability of 1/2. In the first case A=X and the other envelope contains 2X and the latter case A=2X and the other envelope contains X. There is no compelling reason to switch.

As before but I open the selected envelope, find that it contains A and then I am invited to switch. I consider that A could come from the envelope pair containing A and 2A or from an envelope pair containing A/2 and A. The first envelope pair occurs with probability P(X=A) and the second envelope pair occurs with probability P(X=A/2). The relative frequency that I have the envelope pair containing A and 2A is P(X=A) / (P(X=A) + P(X=A/2)) and the relative frequency that I have the envelope pair containing A/2 and A is P(X=A/2) / (P(X=A) + P(X=A/2)). By switching I will gain A with a probability of is P(X=A) / (P(X=A) + P(X=A/2)) or lose A/2 with a probability of is P(X=A/2) / (P(X=A) + P(X=A/2)). I will choose to switch if my expected gain exceeds my expected lose i.e. if

A * P(X=A) / (P(X=A) + P(X=A/2)) > A/2 * P(X=A/2) / (P(X=A) + P(X=A/2))

Since all the terms are positive, divide both sides by A and multiply both sides by (P(X=A) + P(X=A/2))

I will choose to switch if P(X=A) > 1/2 * P(X=A/2).

--Norman Sheldon (talk) 10:44, 12 September 2012 (UTC)

The not switching argument

The not switching argument: Now suppose you reason as follows:

1. I denote by B the amount in the envelope that I have not selected. 2. The probability that B is the smaller amount is 1/2, and that it is the larger amount is also 1/2. 3. The envelope I selected may contain either 2B or B/2. 4. If B is the smaller amount, then the selected envelope contains 2B. 5. If B is the larger amount, then the selected envelope contains B/2. 6. Thus the selected envelope contains 2B with probability 1/2 and B/2 with probability 1/2. 7. So the expected value of the money in the selected envelope is

   ½ (2B) + ½ (B/2) = 5B/4

8. This is greater than B, so I gain on average by not swapping.

Hence, if I know the content of the other envelope then I will not switch; and this is true whatever the value is of the other envelope.

We now have the contradiction that if we assign a value of A to the contents of the selected envelope then we will gain by switching but if we assign a value of B to the contents of the other envelope then we will lose by switching. So what is going on here?

We have two scenarios (call them Game A where we assign A to the contents of the selected envelope and Game B where we assign a value of B to the contents of the other envelope). In Game A we hold the value A and can either gain A or lose A/2 each with a probability of ½. So the odds being offered are 2 to 1 compared to an evens risk, this is very attractive to anyone. In Game B we hold either 2B or B/2 and can either lose B or gain B/2. So the odds being offered are 1 to 2 compared to an evens risk, this is unattractive even to an inveterate gambler. These two scenarios cannot co-exist, so when we say assign A to the amount in the selected envelope we are correct half the time (and ditto for assigning B to the other envelope).

Still not clear? Okay, suppose you know that the selected envelope contains A and the other envelope contains B, without using the words “if”, “whether”, “else” or “otherwise”, please answer do you want to switch?

To add a little to Donald Rumsfeld’s quote, there are also unknown knowns, i.e. we can assign known amounts A and B but we still do not know what they are. Until both envelopes are opened, we do not know (especially for a one-off situation) if we gain or lose by switching. Welcome to the quantum world.

--Norman Sheldon (talk) 18:27, 8 October 2012 (UTC)

The symbol A stands for different things???

One sentence in a proposed "resolution" of the paradox in the article reads as follows:

""'A common way to resolve the paradox, both in popular literature and in the academic literature in philosophy, is to observe that A stands for different things at different places in the expected value calculation, step 7 above."

This is complete nonsense. In all cases, A stands for the unknown amount of money in the first envelope selected before deciding whether to switch.

It is just like saying that a random variable that takes the values 1 and 2 with equal probabilities, both equal to 1/2, stands for "different things" depending on whether it equals 1 or equals 2.Daqu (talk) 20:02, 23 February 2013 (UTC)

Daqu, I tend to agree with you but the real problem is that the problem itself is not that well defined, it is a self-inflicted injury. In other words there is no puzzle initially, it is quite obvious that switching is not advantageous (certainly in the real world), the puzzle is created by proposing a bogus line of argument. The exact solution depends on exactly what the bogus line of argument is taken to be. Martin Hogbin (talk) 09:39, 8 March 2013 (UTC)

My solution @zazaalaza

Firstly I am offered the possibility to choose between switching and not switching. The problem in the line of reasoning above is that after line 8. the logic jumps directly to after the switch. After calculating that i gain more, why can't I calculate further? As I know that the opposite of the expected value is the expected loss, I want to calculate that too before switching. In other worlds, I want to calculate the expected value of the second possible action: I do not switch. Following this logic I can make the exact same calculation by denoting the other envelope with B and come to the realization that the expected loss equals 5/4. So if EV = 5/4 and EL=5/4 there is no reason to switch because even if i gain more, in the same time i lose more too.

Or in other words: expected value of switching = expected value of not switching — Preceding unsigned comment added by 5.15.29.71 (talk) 14:18, 13 March 2013 (UTC)

edit: it doesn't matter what the value of action A is untill it is equal to the value of action B — Preceding unsigned comment added by 5.15.29.71 (talk) 14:30, 13 March 2013 (UTC)

Flawed Logic

I don't understand why this paradox is such a big deal. I see this problem differently than how I understand the article. As I understand the problem, you are given a random Boolean value that you aren't allowed to check. You are then asked if you want to negate that Boolean value. Even if you can check only one of the envelopes, the amount of money contained within either envelope, whether measured relatively or absolutely, is not dependent upon the probability that it is the greater value. It is simply the amount of risk involved in switching. Another way of phrasing this problem would be to flip a coin and call heads or tails, then, without knowing the outcome, be given the ability to change your prediction. Without any knowledge of the outcome, you cannot make any sort of educated guess about switching; blindly changing whether you guessed favorably or not. If this logic is explained in the article, I haven't properly understood. I would subsequently like to be pointed to the section that states this and some articles be recommended to enhance my knowledge and understanding. Googol30 (talk) 01:42, 22 May 2013 (UTC)

I tend to agree with you. This paradox is essentially a self-inflicted injury. To most people it is obvious that swapping has no advantage and this does indeed turn out to be correct.
It is, however, possible to propose a line of argument in which there does appear to be an advantage in swapping. The problem is in finding the flaw in the given argument. The exact location and nature of the flaw in the proposed line of argument depends on exactly what the argument is. As expected, whenever the proposed argument for swapping is clearly defined, the flaw in it becomes apparent. Martin Hogbin (talk) 08:10, 22 May 2013 (UTC)
In general, I agree with both of you. The issue is less that the problem itself is difficult to grasp, and more that a "standard" way of examining the problem yields results that are clearly incorrect. Hence, we should be concerned about using those methods at least some of the time because we might unintentionally repeat the error in a way that would be less easy to catch. Dlepzelter (talk) 19:30, 31 October 2013 (UTC)

Odd Solution?

It would seem that with all of those multitudes of words spent on the topic, no-one came up with the obvious: you should always switch envelopes if the amount contained within your envelope is odd and never if it is possible for the amount to be odd, but you find an even quantity.70.90.204.42 (talk) 16:13, 1 May 2014 (UTC)

Unfortunately, as far as I know, no one has yet published such a solution. So wikipedia can't cite it. Richard Gill (talk) 16:14, 1 June 2014 (UTC)
That is correct only if we assume integer amounts. The general case is that the amounts can be continuous, so this distinction cannot be made. --Pan-wiki-tsik (talk) 18:21, 4 June 2014 (UTC)

Rationalization to the non-probabilistic approach

Labeling both envelopes (such as A and A/2 or X and 2X) is faulty because it presumes knowledge of both envelopes which, while adhering to pure logic, is not sound since logically the universe consists of knowledge of only 1 envelope. That said, knowing the contents of 1 envelope marks its contents as A (vs. 2A or A/2) rather than X or 2X (or A or A/2). This makes the first assertion true: either one gains A or loses A/2, in which case one should always choose the other envelope. It results in the same infinite loop as the probabilistic argument.

Of the 2 approaches, this is not a unique paradox. It is equivalent to the St. Petersburg paradox of infinite expected value. But I have to say, the expected value of a discrete sample space being the Euclidean mean of the values of that sample space is a somewhat ludicrous concept. One cannot *expect* a sum of money to be (2A+A/2)/2 = 1.25*A because it's not one of the options! I admit this flies against modern probability theory, but in the case of resolving a paradox it becomes reasonable.

The problem as it exists is faulty because allowing the player to make an infinite amount of decisions will result in exactly that; if the player is only allowed to make 1 further decision after his/her first (e.g. in the Monty Hall problem), s/he should opt for the other envelope regardless. — Preceding unsigned comment added by 173.168.18.15 (talk) 20:50, 8 June 2014 (UTC)

You are fully correct, IP 173.168.18.15, because the nonsensical term  (2A+A/2)/2=1.25*A  does a priory imply  "A≠A".  So a priori putting up with "2A"=2(A/2) and "A/2"=(2A)/2). Consequently you have to solve this nonsensical term as follows: "The expected value is (A+A)/2". Gerhardvalentin (talk) 07:33, 16 June 2014 (UTC)

Ali Baba Variant

The cited Nalebuff article lists multiple possible answers to the apparent paradox. The only one which implies either Ali or Baba should possibly want to trade is one in which there is a maximum apparent amount of money that might be in the envelope (Nalebuff, The Other Person's Envelope is Always Greener, 1989, pp. 175-177).209.6.139.197 (talk) 04:44, 17 October 2013 (UTC)David Lepzelter

Nalebuff shows the only scenario where the expected value of (2A+A/2)/2=1.25A can ever apply: If A is the "initial amount" contained in the first envelope (say A), and then the second envelope (say B) is equally likely to be filled – thereafter – with either 2A or A/2. Yes, if Ali got the "first" envelope, then he should switch and can expect to get 1,25*A. Whereas Baba, by switching must expect to get only 0.8*B.

But as, in general, "any" envelope is equally likely to having been picked first, the term (2A+A/2)/2 is nonsensical, because from the outset this implies  "A≠A". So a priori putting up with "2A"=2(A/2) and "A/2"=(2A)/2). Consequently you have to solve this nonsensical term as follows: "The expected value is (A+A)/2". Gerhardvalentin (talk) 08:22, 16 June 2014 (UTC)

What am I missing?

Surely any actual playable instance of this game has an underlying probability distribution of the amounts in the envelopes? And, because it has to add up to 1 over all amounts, it has to converge to 0 as the amounts get ever higher. So, given this probability distribution, and a value-in-the-first-envelope A, the probability that A is the smaller envelope cannot be exactly 1/2 over all possible values of A.

Consider:

  • You play the game 1,000,000 times. You don't swap, just take the amount in the first envelope. You write down the highest amount you've seen so far.
  • Now, on the 1,000,001th go, you're offered an amount higher than any you've previously seen. Intuitively, would you swap? — Preceding unsigned comment added by 217.150.97.26 (talk) 12:56, 15 October 2013 (UTC)
I think you've hit the nail on the head, the original argument to support switching assumes that 2X is equally as likely as (1/2)X, and that implies either an upper limit (and we know the maths don't work then, because with a limit of $100 you never swap $100) or constant probability over 0 to infinity with the implication that the expected winnings are infinite, or tend to infinity as we increase the value of the game limit towards infinity. Gomez2002 (talk) 12:39, 6 August 2014 (UTC)
Yes I agree. Either there is an upper limit, in which case it is easy to show that on average you do not gain by swapping, or there is not, in which case the expected value in both envelopes is infinite. I think this is the conclusion in the literature when you read it. Martin Hogbin (talk) 13:24, 6 August 2014 (UTC)

Randomized solutions

There's a basically equivalent but simpler formulation than the one in the main article. Select a strictly decreasing function, f(x), defined on the non-negative real numbers, which is 1 at 0 and approaches 0 as x approaches infinity; for instance, exp(-x), 1/(1 + x), 1/log(e + x), etc. If the originally chosen envelope contains an amount A, swap with probability f(A). For any two fixed amounts, A0 < A1, if the smaller amount is chosen, it will be more likely swapped than if the larger amount is chosen. The expected value is (A0 + A1)/2 + (f(A0) - f(A1))(A1 - A0)/2. If the amounts are A and 2A, the expected value is 3A/2 + (f(A) - f(2A))A/2— Preceding unsigned comment added by Thinkatron (talkcontribs) 06:46, 14 August 2014 (UTC)