[ExI] Strong AI Hypothesis: logically flawed?
bbenzai at yahoo.com
Thu Oct 2 17:15:58 UTC 2014
William Flynn Wallace <foozler83 at gmail.com> wrote:
>>?? Suppose you are informed that
>>you have a disease that causes you to die whenever you fall asleep at
>>night so that the person who wakes up in the morning is a completely
>>different Dan Ust who shares your memories. This has been happening
>>every day of your life, but you have only just found out about it.
>>Would this information worry you or make any difference to how you
>>live your life?
>Absolutely YES. It means that someone with my body and memories will
>awake tomorrow but it will not be me. It means that I will die today.
I think it means that we need to look very carefully at what we mean by
'death'. Let's say that instead of once a day, we all 'die' ten million
trillion trillion times every second, and are replaced by "someone else
with..", etc. How can that possibly make any difference to you? What
does that actually mean for the concept of 'you'?
The point of the argument is to show that /it doesn't matter/. In fact,
I think there's a bit of a trick in the statement "a completely
different Dan Ust who shares your memories". That's a bit like saying
"a completely different tune that shares all the same notes, in the same
order, with the same timing". What we're talking about with uploading
or copying or whatever, is equivalent to the same tune being played on a
different instrument. Calling it a different tune is clearly incorrect.
More information about the extropy-chat