<div dir="ltr"><div dir="ltr"><br></div><br><div class="gmail_quote gmail_quote_container"><div dir="ltr" class="gmail_attr">On Sat, May 9, 2026 at 4:42 AM John Clark via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org">extropy-chat@lists.extropy.org</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr"><div dir="ltr"><div dir="ltr"><font size="4" face="tahoma, sans-serif"><b>David Chalmers wanted to know why the brain’s behavior is accompanied by experience at all, but it seems to me that if Charles Darwin was right then there can be only one answer to that question, because consciousness is the inevitable byproduct of intelligence. That's because Natural Selection can't directly detect consciousness any better than we can detect it in other people, and nothing can select for something that it can't see. But <span class="gmail_default" style="font-family:arial,helvetica,sans-serif">N</span>atural <span class="gmail_default" style="font-family:arial,helvetica,sans-serif">S</span>election <span class="gmail_default" style="font-family:arial,helvetica,sans-serif">CAN</span> see intelligent behavior. </b></font></div></div></div></blockquote><div><br></div><div>We've probably had this discussion before, but I disagree that consciousness is the inevitable byproduct of intelligence. There are clearly multiple ways to achieve intelligence. abstract discrete logic gates and running on physical phenomenal qualities being two very different types. If one is more efficient than the other, evolution will naturally select for that.</div><div><br></div><div>Again, my prediction is that there are far more efficient ways to achieve intelligence than using abstract, brute force, discrete logic gates.</div><div><br></div><div>Just because something can see colored light better than you can, doesn't mean its redness will be anything like yours. You always seem to ignore the fact that the word 'red' is nothing like a redness quality. A redness quality will never be the byproduct of an abstract word like red.</div><div><br></div><div><br></div><div><br></div><div><br></div><div> </div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr"><div dir="ltr"><div dir="ltr"><font size="4" face="tahoma, sans-serif"><b><br></b></font></div><div dir="ltr"><font size="4" face="tahoma, sans-serif"><b>Chalmers also claimed that even after hypothetically accounting for our entire behavior, and for all our reports about our inner life, there would still be an explanatory gap between brain processes and experience; and I agree with Chalmers about that because any iterative sequence of "why" questions will either go on forever or terminate in a brute fact, such as consciousness is the way data feels when it is being processed intelligently. </b></font></div><div dir="ltr"><font size="4" face="tahoma, sans-serif"><b><br></b></font></div><div><font size="4" face="tahoma, sans-serif"><b><span class="gmail_default">John K Clark</span><br></b></font></div><div class="gmail_quote"><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><br>
</blockquote></div></div>
</div>
_______________________________________________<br>
extropy-chat mailing list<br>
<a href="mailto:extropy-chat@lists.extropy.org" target="_blank">extropy-chat@lists.extropy.org</a><br>
<a href="http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat" rel="noreferrer" target="_blank">http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat</a><br>
</blockquote></div></div>