1 of 20

Models of decision-making

based on logical counterfactuals

Vladimir Slepnev

2 of 20

Goals

Modeling human decision-making is not really a goal for me.

Neither is building efficient AI algorithms.

What is the “correct” decision, never mind the reasoning cost?

What is the “correct” way of decision-making that we’re trying to approximate?

3 of 20

Overview

  • Describe a simple decision problem
  • Solve it in an overcomplicated way
  • Generalize the approach
  • Solve some more problems
  • Give an outline of further research

4 of 20

A simple decision problem

“Would you like some chocolate?”

  • Yes → you get some chocolate.
  • No → you don’t.

5 of 20

A simple decision problem

“Would you like some chocolate?”

  • Yes → you get some chocolate.
  • No → you don’t.

Desiderata for a model:

  • Two mathematical objects: U (universe) and A (agent).
  • Both U and A should be “completely deterministic”.
  • The description of U should “contain” the description of A.
  • The descriptions of both U and A should be “completely known” to A.
  • A’s decision should be based on “reasoning” about U and A.

6 of 20

Our proposed model

  • U and A are sentences in Peano arithmetic (PA) without free variables.
  • The truth value of A indicates whether the agent says "yes" or "no".
  • The truth value of U indicates whether the agent gets chocolate or not.

7 of 20

Our proposed model

  • U and A are sentences in Peano arithmetic (PA) without free variables.
  • The truth value of A indicates whether the agent says "yes" or "no".
  • The truth value of U indicates whether the agent gets chocolate or not.

A mutually recursive definition of U and A:

“If you say "yes", I will give you chocolate, otherwise I won't.”

“If I can prove that saying “yes” leads to chocolate, then I say “yes”, otherwise “no”.”

All self-references occur within Gödel number quotes, therefore such U and A exist, by the Diagonal Lemma.

8 of 20

Analysis

It’s easy to prove that U and A are both true.

9 of 20

Analysis

It’s easy to prove that U and A are both true.

What if we changed the problem a little? Reward “no” with chocolate:

Now A is false (as long as PA is consistent), and U is again true.

It feels like A is trying to make U true, in order to get some chocolate :-)

10 of 20

But does it generalize?

  • Many possible outcomes
  • Many possible actions
  • Many possible worlds
  • Probabilistic strategies
  • Reacting to observations
  • Multiple instances of yourself
  • Multiple competing agents
  • Exotic kinds of uncertainty
  • ...

11 of 20

Newcomb’s problem

  • There are two closed boxes in front of me.
  • I can take either box 1 and box 2 (“two-boxing”), or only box 2 (“one-boxing”).
  • Before the experiment, a perfect predictor predicted my action.
  • The prediction was used to fill the boxes.
  • Box 1 always contains $1000.
  • Box 2 contains $1000000 iff the predictor predicted that I would one-box.

12 of 20

Newcomb’s problem

We will define these sentences in PA:

  • A is true iff the agent one-boxes.
  • P is true iff the predictor predicted that the agent would one-box.
  • B1 is true iff the agent gets the $1000 from box 1.
  • B2 is true iff the agent gets the $1000000 from box 2.

We will use these equations:

  • P ↔ A
  • B1 ↔ ¬A
  • B2 ↔ P
  • A ↔ ?

13 of 20

Newcomb’s problem

  • P ↔ A

“The predictor predicts that I one-box iff I actually one-box.”

  • B1 ↔ ¬A

“I get the contents of the first box iff I two-box.”

  • B2 ↔ P

“I get the contents of the second box iff the predictor predicted that I would one-box.”

  • A ↔ ?

“If I can get the contents of both boxes by one-boxing, then I one-box;

otherwise, if I can get both boxes by two-boxing, then I two-box;

otherwise, if I can get only box 2 by one-boxing, then I one-box;

otherwise, if I can get only box 2 by two-boxing, then I two-box;

otherwise, if I can get only box 1 by one-boxing, then I one-box;

otherwise I two-box.”

14 of 20

Newcomb’s problem

The completed equations:

It’s easy to prove that A is true, B1 is false, and B2 is true.

Thus, our approach favors one-boxing.

15 of 20

Absent-minded driver problem

  • To get home from work, you need to pass two identical intersections.
  • At each intersection you can either continue or exit.
  • You need to continue at the first intersection, and exit at the second.
  • You’re absent-minded and can’t remember which intersection you’re at.
  • To allow probabilistic choices, you observe a coin flip at each intersection.
  • What strategy gives you the best chance of getting home?

(Slightly modified from Piccione and Rubinstein, 1997)

16 of 20

Absent-minded driver problem

We will define these sentences in PA:

  • A1 is true iff you continue in case of heads
  • A2 is true iff you continue in case of tails
  • U11 is true iff you get home in case of (heads, heads)
  • Similar for U12, U21, U22

U11 ↔ U22 ↔ ⊥

U12 ↔ A1∧¬A2

U21 ↔ ¬A1∧A2

A1 ↔ ?

A2 ↔ ?

17 of 20

Absent-minded driver problem

A1 ↔ ?

A2 ↔ ?

“If A1∧A2 provably implies that all Uij are true, then make A1 and A2 true;

otherwise, if A1∧¬A2 provably implies that all Uij are true, then make A1 true and A2 false;

{...}

otherwise, if A1∧A2 provably implies that exactly three of Uij are true, then make A1 and A2 true;

{…}”

The equations begin like this:

18 of 20

Other proposed models

Using Gödel-Löb provability logic instead of PA:

  • Use ◻ instead of Prov
  • Use modal fixed points instead of the Diagonal Lemma
  • Equivalent to the PA approach, because GL is adequate for PA (Solovay)
  • Decidable!

Using computer programs that look for proofs, instead of arithmetic formulas:

  • Chronologically, the first approach we came up with
  • If programs have access to provability oracles, this is also equivalent to PA
  • If programs enumerate proofs up to a fixed size, it’s “almost” equivalent
  • Undecidable in general

19 of 20

Further work

From decision theory to game theory

  • What if there are multiple agents proving things about each other?
  • What do you want other agents to prove about you?
  • How does “proof warfare” influence cooperation, bargaining, blackmail...

From perfect certainty to uncertainty

  • How do you handle uncertainty about mathematical facts?
  • How do you handle uncertainty about your self-description?
  • How do you handle uncertainty about your values?

20 of 20

Questions?