[extropy-chat] How to be copied into the future?

Lee Corbin lcorbin at rawbw.com
Mon Apr 23 03:17:16 UTC 2007

Mike writes

> Samantha wrote:
>> Lee Corbin wrote:
>> > That's a truism.  But I take you to mean that you'd be willing to
>> > defer to its terms.  No?  As for me, if it is cheap in terms of time
>> > and resources, and the advanced entities could save us---but
>> > choose not to (for whatever reasons)---then I'd just as soon
>> > they not exist either.  The bastards.
>> You would prefer no intelligence at all in the local universe if that
>> intelligence wasn't human or formerly human?

That's right.  At least under the condition that they could have saved
us---at no expense yet!---but did not deign to.  To hell with them.

You can't let people push you around.

> If They are so full of themselves that they can't be bothered to run
> LeeCorbin as a screen saver, then Lee is justifiably bitter

It's the least they could do.  And I can't believe that not everyone would
be a bit miffed at their actions:

Suppose an AI takes over the world, answers all our math, physics,
and general knowledge questions, and then suddenly says "Oh, by the
way, I've decided I don't want you around no more. Tuesday---then
that's it!"

> and is (verbally) spitting on them.  :)

Yeah---and don't forget I'm making a fuss for the rest of you too!
Let's plant fear right now:  fear that it's all a test and that we have a 
secret way to pull the plug---or that they're in a simulation and we're
waiting to see if they pass a test of basic gratitude!  :-)


More information about the extropy-chat mailing list