[ExI] A new theory of consciousness: conditionalism
Jason Resch
jasonresch at gmail.com
Fri Aug 25 19:11:30 UTC 2023
Hi Adrian,
Thanks for replying.
On Fri, Aug 25, 2023, 2:55 PM Adrian Tymes via extropy-chat <
extropy-chat at lists.extropy.org> wrote:
> Consciousness might include conditionalism, but would it not also include
> the ability to formulate new conditionals and evaluate whether they match
> perceived reality? I do not see that capability within your definition of
> conditionalism.
>
I see the divide between those things as the divide between the mind (as an
adaptive, learning, process over time) vs. any single "thought-moment" (a
single self-isolated state of consciousness).
To the mind I would ascribe the formulation of new states of consciousness,
but any particular state of consciousness (under conditionalism) is that
knowledge state embodied by a conditional relation.
For the mind to make the decisions, that is, to act on environmental or
internal information, it must use conditionals, and these conditional
states are (what I propose) what create/invoke/are the knowledge states
that we associate with consciousness.
Jason
> On Fri, Aug 25, 2023 at 10:49 AM Jason Resch via extropy-chat <
> extropy-chat at lists.extropy.org> wrote:
>
>> I would like to propose a theory of consciousness which I think might
>> have some merit, but more importantly I would like to see what criticism
>> others might have for it.
>>
>> I have chosen the name "conditionalism" for this theory, as it is based
>> loosely on the notion of conditional statements as they appear in both
>> regular language, mathematics, and programming languages.
>>
>> At a high level, states of consciousness are states of knowledge, and
>> knowledge is embodied by the existence of some relation to some truth.
>>
>> A conditional is a means by which a system can enter/reach a state of
>> knowledge (i.e. a state of consciousness) if and only if some fact is true.
>> A simple example using a programming language:
>>
>> if (x >= 5) then {
>> // knowledge state of x being greater than or equal to 5
>> }
>>
>> I think this way of considering consciousness, as that existing between
>> those two braces: { } can explain a lot.
>>
>> 1. Consciousness is revealed as an immaterial, ephemeral relation, not
>> any particular physical thing we can point at or hold.
>>
>> 2. It provides for a straight-forward way to bind complex states of
>> consciousness, though conjunction, for example:
>> If (a and b) {
>> // knowledge of the simultaneous truth of both a and b
>> }
>> This allows states of consciousness to be arbitrarily complex and varied.
>>
>> 3. It explains the causal efficacy of states of consciousness. All we
>> need to do is link some action to a state of knowledge. Consciousness is
>> then seen as antecedent to, and a prerequisite for, any intelligent
>> behavior. For example:
>> If (light == color.red) {
>> slowDown();
>> }
>>
>> 4. It shows the close relationship between consciousness and information,
>> where information is defined as "a difference that makes a difference", as
>> conditionals are all about what differences make which differences.
>>
>> 5. It shows a close relationship between consciousness and
>> computationalism, since computations are all about counterfactual and
>> conditional relations.
>>
>> 6. It is also supportive of functionalism and it's multiple
>> realizability, as there are many possibile physical arrangements that lead
>> to conditionals.
>>
>> 7. It's clear there neural networks firings is all about conditionals and
>> combining them in whether or not a neuron will fire and which other neurons
>> have fired binds up many conditional relations into one larger one.
>>
>> 8. It seems no intelligent (reactive, deliberative, contemplative,
>> reflective, etc.) process can be made that does not contain at least some
>> conditionals. As without them, there can be no responsiveness. This
>> explains the biological necessity to evolve conditionals and apply them in
>> the guidance of behavior. In other words, consciousness (states of
>> knowledge) would be strictly necessary for intelligence to evolve.
>>
>>
>> Jason
>> _______________________________________________
>> extropy-chat mailing list
>> extropy-chat at lists.extropy.org
>> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>>
> _______________________________________________
> extropy-chat mailing list
> extropy-chat at lists.extropy.org
> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20230825/a0a4b7f9/attachment-0001.htm>
More information about the extropy-chat
mailing list