diff options
author | Nicholas Johnson <nick@nicholasjohnson.ch> | 2024-05-27 00:00:00 +0000 |
---|---|---|
committer | Nicholas Johnson <nick@nicholasjohnson.ch> | 2024-05-27 00:00:00 +0000 |
commit | 628046738b0e4f410c639dd4844925ff044c79d2fb14b0e42722f1bee733f1ad (patch) | |
tree | cc1af60eedfa34aca0c24a6f1f6edfc554b6912715dc090bc8f124527e857caf /content/entry/newcombs-paradox-resolved.md | |
parent | 46e98fe4f8c4c373ccb42427122f1fe032cc68038ec3e13dcf43dec31b874a8a (diff) | |
download | journal-628046738b0e4f410c639dd4844925ff044c79d2fb14b0e42722f1bee733f1ad.tar.gz journal-628046738b0e4f410c639dd4844925ff044c79d2fb14b0e42722f1bee733f1ad.zip |
Fix tons of links
Diffstat (limited to 'content/entry/newcombs-paradox-resolved.md')
-rw-r--r-- | content/entry/newcombs-paradox-resolved.md | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/content/entry/newcombs-paradox-resolved.md b/content/entry/newcombs-paradox-resolved.md index 1c5d3b2..7140b9e 100644 --- a/content/entry/newcombs-paradox-resolved.md +++ b/content/entry/newcombs-paradox-resolved.md @@ -4,7 +4,7 @@ date: 2020-11-28T00:00:00 draft: false --- # Background -I "solved" [Newcomb's Paradox](https://www.wikipedia.org/wiki/Newcomb%27s_paradox) about 3 years ago if I remember right. I use solved in quotes because you don't really "solve" a paradox. Paradoxes only seem to be contradictory at first glance. But, upon further inspection, they lead you to a new understanding of the problem where the paradox disappears. In other words, paradoxes arise out of a flawed or incomplete perspective. Newcomb's paradox in particular arises out of a misunderstanding of [free will](/2020/06/19/free-will-is-incoherent-part-1). +I "solved" [Newcomb's Paradox](https://en.wikipedia.org/wiki/Newcomb%27s_paradox) about 3 years ago if I remember right. I use solved in quotes because you don't really "solve" a paradox. Paradoxes only seem to be contradictory at first glance. But, upon further inspection, they lead you to a new understanding of the problem where the paradox disappears. In other words, paradoxes arise out of a flawed or incomplete perspective. Newcomb's paradox in particular arises out of a misunderstanding of [free will](/2020/06/19/free-will-is-incoherent-part-1/). # Telescoping Method Before I give away the "solution" to Newcomb's Paradox, I want to talk about my method for solving philosophical problems in general. I call it the telescoping method. When I am confronted with a philosophical problem, the first thing I do is try to understand the essence of the problem. To do that, I look at which abstractions the problem uses. I break them down and down until the problem makes sense. Then, one by one, I build up the abstractions again so I can explain in words what the solution is. I've found it to be very effective for philosophy. I'll use Newcomb's Paradox to show how it works. @@ -21,11 +21,11 @@ Here's the problem from Wikipedia ([CC BY-SA 3.0](https://creativecommons.org/li > The player does not know what the predictor predicted or what box B contains while making the choice. ## The Paradox -Let's make it clear why this situation is a paradox before we attempt to resolve it. We are going to assume that the player is trying to maximize their profits. There are 2 strategies from [game theory](https://www.wikipedia.org/wiki/Game_theory) that you can apply to Newcomb's Paradox. +Let's make it clear why this situation is a paradox before we attempt to resolve it. We are going to assume that the player is trying to maximize their profits. There are 2 strategies from [game theory](https://en.wikipedia.org/wiki/Game_theory) that you can apply to Newcomb's Paradox. -The first is the [expected utility hypothesis](https://www.wikipedia.org/wiki/Expected_utility_hypothesis). It says you should take only box B. Its reasoning goes like this: Imagine you watched 1,000 philosophers play. Philosophers are split 50/50 on the issue, so about 500 would pick only box B, winning $1,000,000 and about 500 would pick both boxes, winning $1000. From a statistical standpoint, it's obvious you should pick box B. +The first is the [expected utility hypothesis](https://en.wikipedia.org/wiki/Expected_utility_hypothesis). It says you should take only box B. Its reasoning goes like this: Imagine you watched 1,000 philosophers play. Philosophers are split 50/50 on the issue, so about 500 would pick only box B, winning $1,000,000 and about 500 would pick both boxes, winning $1000. From a statistical standpoint, it's obvious you should pick box B. -The second is the [strategic dominance principle](https://www.wikipedia.org/wiki/Strategic_dominance). It says you should take boxes A and B. Its reasoning goes like this: After the predictor has made its prediction and either put the $1,000,000 in box B or not, it can't change the amounts in the boxes. Therefore, choosing both boxes will always yield $1000 more than choosing only box B. You should take both boxes. +The second is the [strategic dominance principle](https://en.wikipedia.org/wiki/Strategic_dominance). It says you should take boxes A and B. Its reasoning goes like this: After the predictor has made its prediction and either put the $1,000,000 in box B or not, it can't change the amounts in the boxes. Therefore, choosing both boxes will always yield $1000 more than choosing only box B. You should take both boxes. At first glance, both of these principles seem very compelling. It's paradoxical because they offer conflicting advice. How can we resolve this? @@ -53,7 +53,7 @@ If it's not the infallible predictor causing the paradox, could the essence of t Even if your sincere intention is to pick only box B before the predictor makes its prediction, yet you change your mind and choose both boxes in the after prediction stage, the predictor will still correctly predict your choice and you will only win $1000. ### Choice -We already examined the infallibility of the predictor and the player intent abstractions. They don't seem to cause the paradox. Perhaps the abstraction of choice is the problem? Newcomb's Paradox assumes it makes sense to talk about the player making a "choice" between 2 boxes and 1 box. But language to describe making a choice between several options is used in plenty of game theory problems. Even though I have shown that [free will is incoherent](/2020/06/19/free-will-is-incoherent-part-1), using what I call "the language of free will" doesn't seem to be an issue for other game theory problems. Why then would it be especially problematic in Newcomb's Paradox? Allow me to defer to some dialogue between Neo and The Oracle from [The Matrix](https://www.wikipedia.org/wiki/The_Matrix): +We already examined the infallibility of the predictor and the player intent abstractions. They don't seem to cause the paradox. Perhaps the abstraction of choice is the problem? Newcomb's Paradox assumes it makes sense to talk about the player making a "choice" between 2 boxes and 1 box. But language to describe making a choice between several options is used in plenty of game theory problems. Even though I have shown that [free will is incoherent](/2020/06/19/free-will-is-incoherent-part-1/), using what I call "the language of free will" doesn't seem to be an issue for other game theory problems. Why then would it be especially problematic in Newcomb's Paradox? Allow me to defer to some dialogue between Neo and The Oracle from [The Matrix](https://en.wikipedia.org/wiki/The_Matrix): > Oracle: Candy? > Neo: Do you already know if I'm going to take it? @@ -68,7 +68,7 @@ TLDR; choose only box B. Long answer: There is a very subtle contradiction in the definition of Newcomb's Paradox. Can you spot it? It says "The player does not know what the predictor predicted or what box B contains while making the choice". The hidden assumption there is that the "choice point" is after the predictor's prediction. This is impossible. The abstraction of choice collapses after the predictor has made the prediction. If we have to pick a point in time where it still makes any sense to talk about a "choice" being made, it would have to be before the predictor made the prediction. The strategic dominance principle is inherently tied to the idea of the player having a free choice after the predictor made the prediction. Therefore, it can't be the solution. -Meanwhile taking only box B is supported by mathematical expected value, which doesn't rely on free choice being available after the prediction. It just says "If you take only box B, you can expect $1,000,000. If you take both boxes, you can expect $1,000". There's no notion of free will there. It's a purely statistical argument. The strategic dominance principle only seems appealing because of the strong intuition of having a free choice after the predictor has made the prediction. While [retrocausality](https://www.wikipedia.org/wiki/Retrocausality) doesn't actually occur in Newcomb's Paradox, it's not a bad mental model for thinking about the problem. Since the predictor is infallible, it has effective retrocausality. What the predictor did in the past is based on the box it already knows you're going to take. There's no real paradox, you just can't outwit the predictor even though your intuitions tell you that you "feel free". +Meanwhile taking only box B is supported by mathematical expected value, which doesn't rely on free choice being available after the prediction. It just says "If you take only box B, you can expect $1,000,000. If you take both boxes, you can expect $1,000". There's no notion of free will there. It's a purely statistical argument. The strategic dominance principle only seems appealing because of the strong intuition of having a free choice after the predictor has made the prediction. While [retrocausality](https://en.wikipedia.org/wiki/Retrocausality) doesn't actually occur in Newcomb's Paradox, it's not a bad mental model for thinking about the problem. Since the predictor is infallible, it has effective retrocausality. What the predictor did in the past is based on the box it already knows you're going to take. There's no real paradox, you just can't outwit the predictor even though your intuitions tell you that you "feel free". You might think it doesn't make sense to prescribe players the strategy of choosing box B only, since they have "already made the choice" whether or not to take only box B. But, consider that by the same token, we have "already made the choice" whether or not to prescribe the player the strategy to take box B. So, it is equally coherent for us to prescribe the player to take box B as it is for the player to actually take box B. Saying there's no point in prescribing the player a course of action is akin to saying you'll just stay in bed all day since you have no free will. The "choice" to do nothing is also not of your own free will. In other words, you're not escaping your lack of free will by doing nothing. We aren't escaping the lack of the player's free will by not prescribing them a best course of action as we don't have free will either. So, there's no reason not to tell the player to take only box B. |