[extropy-chat] Re: Overconfidence and meta-rationality

Eliezer S. Yudkowsky sentience at pobox.com
Sun Mar 20 06:25:06 UTC 2005


Robin Hanson wrote:
> At 12:57 AM 3/13/2005, Eliezer S. Yudkowsky wrote:
> 
>> If I had to select out two points as most important, they would be:
>> 1) Just because perfect Bayesians, or even certain formally imperfect 
>> Bayesians that are still not like humans, *will* always agree; it does 
>> not follow that a human rationalist can obtain a higher Bayesian score 
>> (truth value), or the maximal humanly feasible score, by deliberately 
>> *trying* to agree more with other humans, even other human rationalists.

>> 2) Just because, if everyone agreed to do X without further argument 
>> or modification (where X is not agreeing to disagree), the average 
>> Bayesian score would increase relative to its current position, it 
>> does not follow that X is the *optimal* strategy.
> 
> These points are stated very weakly, basically just inviting me to 
> *prove* my claims with mathematical precision.  I may yet rise to that 
> challenge when I get more back into this.

We are at odds about what the math here actually *says*.  I don't regard 
the sequiturs (the points above where I say, "it does not follow") as 
things that are a trivial distance from previously accomplished math. 
They seem to me almost wholly unrelated to all work on Aumann Agreement 
Theorems done so far.

Since state information is irrelevant to our dispute, it would seem that 
we disagree about the results of a computation.

Here's at least one semisolid mathematical result that I scribbled down 
in a couple of lines of calculus left to the reader, a result which I 
intuitively expected to find, and which I would have found to be a much 
more compelling argument in 2003.  It is that, when two Bayesianitarians 
disagree about probabilities P and ~P, they can always immediately 
improve the expectation of the sum of their Bayesian scores by averaging 
together their probability estimates for P and ~P, *regardless of the 
real value*.

Let the average probability equal X.  Let the individual pre-averaging 
probabilities equal X+d and X-d.  Let the actual objective frequency 
equal P.  The function:

f(d) = p*[log(x+d) + log(x-d)] + (1-p)*[log(1-x-d) + log(1-x+d)]

has a maximum at d=0, regardless of the value of p.  f'(0)=0 and f''(0) 
is negative.  If my math hasn't misled me there's a couple of other 
points where f'(d)=0 but they're presumably inflection points or minima 
or some such.  I didn't bother checking which is why I call this a 
semisolid result.

Therefore, if two Bayesianitarian *altruists* find that they disagree, 
and they have no better algorithm to resolve their disagreement, they 
should immediately average together their probability estimates.

I would have found this argument compelling in 2003 because at that 
time, I was thinking in terms of a "Categorical Imperative" foundation 
for probability theory, i.e., a rule that, if all observers follow it, 
will maximize their collective Bayesian score.  I thought this solved 
some anthropic problems, but I was mistaken, though it did shed light. 
Never mind, long story.

To try and translate my problem with my former foundation without going 
into a full-blown lecture on the Way:  Suppose that a creationist comes 
to me and is genuinely willing to update his belief in evolution from ~0 
to .5, providing that I update my belief from ~1 to .5.  This will 
necessarily improve the expectation of the sum of our Bayesian scores.

But:

1)  I can't just change my beliefs any time I please.  I can't cause 
myself not to believe in evolution by an act of will.  I can't look up 
at a blue sky and believe it to be green.  I account this a strength of 
a rationalist.
2)  Evolution is still correct regardless of what the two of us do about 
our probability assignments.  I would need to update my belief because 
of a cause that I believe to be uncorrelated with the state of the 
actual world.
3)  Just before I make my Bayesianitarian act of sacrifice, I will know 
even as I do so, correctly and rationally, that evolution is true.  And 
afterward I'll still know, deep down, whatever my lips say...
4)  I have other beliefs about biology that would be inconsistent with a 
probability assignment to evolution of 0.5.
5)  I do wish to be an altruist, but in the service of that wish, it is 
rather more important that I get my beliefs right about evolution than 
that J. Random Creationist do so, because JRC is less likely to like 
blow up the world an' stuff if he gets his cognitive science wrong.

If I were an utterly selfish Bayesianitarian, how would I maximize only 
my own Bayesian score?  It is this question that is the foundation of 
*probability* theory, the what-we-deep-down-know-will-happen-next, 
whatever decisions we make in the service of altruistic utilities.

Point (2) from my previous post should now also be clearer:

 >> 2) Just because, if everyone agreed to do X without further argument
 >> or modification (where X is not agreeing to disagree), the average
 >> Bayesian score would increase relative to its current position, it
 >> does not follow that X is the *optimal* strategy.

Two Aumann agents, to whom other agents' probability estimates are 
highly informative with respect to the state of the actual world, could 
also theoretically follow the average-together-probability-estimates 
algorithm and thereby, yes, improve their summed Bayesian scores from 
the former status quo - but they would do worse that way than by 
following the actual Aumann rules, which do not lend credence as such to 
the other agent's beliefs, but simply treat those beliefs as another 
kind of Bayesian signal about the state of the actual world.

Thus, if you want to claim a mathematical result about an expected 
individual benefit (let alone optimality!) for rationalists deliberately 
*trying* to agree with each other, I think you need to specify what 
algorithm they should follow to agreement - Aumann agent? 
Bayesianitarian altruist?  In the absence of any specification of how 
rationalists try to agree with each other, I don't see how you could 
prove this would be an expected individual improvement.  Setting both 
your probability estimates to zero is an algorithm for trying to agree, 
but it will not improve your Bayesian score.

-- 
Eliezer S. Yudkowsky                          http://singinst.org/
Research Fellow, Singularity Institute for Artificial Intelligence



More information about the extropy-chat mailing list