In decision theory, the sure-thing principle states that a decision maker who decided they would take a certain action in the case that event E has occurred, as well as in the case that the negation of E has occurred, should also take that same action if they know nothing about E.
The principle was coined by L.J. Savage:[1]
A businessman contemplates buying a certain piece of property. He considers the outcome of the next presidential election relevant. So, to clarify the matter to himself, he asks whether he would buy if he knew that the Democratic candidate were going to win, and decides that he would. Similarly, he considers whether he would buy if he knew that the Republican candidate were going to win, and again finds that he would. Seeing that he would buy in either event, he decides that he should buy, even though he does not know which event obtains, or will obtain, as we would ordinarily say. It is all too seldom that a decision can be arrived at on the basis of this principle, but except possibly for the assumption of simple ordering, I know of no other extralogical principle governing decisions that finds such ready acceptance.
— p. 21
Savage formulated the principle as a dominance principle, but it can also be framed probabilistically.[2] Richard Jeffrey[2] and later Judea Pearl[3] showed that Savage's principle is only valid when the probability of the event considered (e.g., the winner of the election) is unaffected by the action (buying the property). Under such conditions, the sure-thing principle is a theorem in the do-calculus[3] (see Bayes networks). Blyth constructed a counterexample to the sure-thing principle using sequential sampling in the context of Simpson's paradox,[4] but this example violates the required action-independence provision.[5]
In the above cited paragraph, Savage illustrated the principle in terms of knowledge. However the formal definition of the principle, known as P2, does not involve knowledge because, in Savage's words, "it would introduce new undefined technical terms referring to knowledge and possibility that would render it mathematically useless without still more postulates governing these terms." Samet[6] provided a formal definition of the principle in terms of knowledge and showed that the impossibility to agree to disagree is a generalization of the sure-thing principle. It is similarly targeted by the Ellsberg and Allais paradoxes, in which actual people's choices seem to violate this principle.[2]
See also
editReferences
edit- ^ Savage, L. J. (1954), The foundations of statistics. John Wiley & Sons Inc., New York.
- ^ a b c Jeffrey, Richard (1982). "The Sure Thing Principle". Proceedings of the Biennial Meeting of the Philosophy of Science Association. 1982 (2): 719–730. doi:10.1086/psaprocbienmeetp.1982.2.192456. JSTOR 192456. S2CID 124506828.
- ^ a b Pearl, Judea (2009). Causality: Models, Reasoning, and Inference (2nd ed.). New York: Cambridge University Press.
- ^ Blyth, C. (1972). "On Simpson's paradox and the sure-thing principle". Journal of the American Statistical Association. 67 (338): 364–366. doi:10.2307/2284382. JSTOR 2284382.
- ^ Pearl, Judea (December 2015). "The sure-thing principle" (PDF). UCLA Cognitive Systems Laboratory, Technical Report R-466.
- ^ Samet, Dov (2022). "The impossibility of agreeing to disagree: An extension of the sure-thing principle". Games and Economic Behavior (338): 390–399. doi:10.1016/j.geb.2022.01.016.