<div dir="ltr"><div dir="ltr"><div class="gmail_default" style="font-family:arial,helvetica,sans-serif"><span style="font-family:Arial,Helvetica,sans-serif">On Thu, Oct 23, 2025 at 9:48 AM Adrian Tymes via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org">extropy-chat@lists.extropy.org</a>> wrote:</span></div></div><div class="gmail_quote gmail_quote_container"><div dir="ltr" class="gmail_attr"><br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><font size="4" face="georgia, serif"><i><span class="gmail_default" style="font-family:arial,helvetica,sans-serif">> </span>I was addressing the terrestrial-scale scenario presented, not<span class="gmail_default" style=""> </span>potential J-Brains (which would occupy different planets entirely).<br></i></font></blockquote><div><br></div><font size="4" face="tahoma, sans-serif"><b>The Guinness Book of World Records no longer recognizes a highest IQ category because of "<i>a lack of a universally agreed-upon standard</i>". It's easy to see why they did that,<span class="gmail_default" style="font-family:arial,helvetica,sans-serif"> t</span>he only one who would have the competence to write a test to find<span class="gmail_default" style="font-family:arial,helvetica,sans-serif"></span> the world's smartest human would be the world's smartest human, and that fact introduces certain obvious<span class="gmail_default" style="font-family:arial,helvetica,sans-serif"> </span>difficulties. </b></font></div><div class="gmail_quote gmail_quote_container"><br></div><font size="4" face="tahoma, sans-serif"><b>How could somebody with just Human intelligence even judge the responses that a superintelligence gave on an IQ test? Suppose the year was 1901 and one of the items on an IQ test was "prove Fermat's Last Theorem" and suppose that somebody had given a proof that was identical to the one that Andrew Wiles gave in 1995, how could anybody know if it was valid? In 1901 even the world's top mathematicians would have had no idea what Wiles was talking about because <span class="gmail_default" style="">in his</span> proof he was using concepts without explanation, he didn't need to because they were common knowledge to all mathematicians in 1995, but were completely unknown to mathematicians in 1901. If Wiles had included all those explanations in his proof then it would've been 10 times as large, and even then it would've probably taken mathematicians at least a decade to fully understand it and realize that Wiles was right.</b></font><div><font face="tahoma, sans-serif" size="4"><b><br></b></font></div><div><font face="tahoma, sans-serif" size="4"><b><span class="gmail_default" style="">John K Clark</span><br></b></font><div><br></div><div><br></div><div><br></div><div><br></div><div> <div class="gmail_quote gmail_quote_container"><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">
<br>
On Thu, Oct 23, 2025 at 9:32 AM John Clark <<a href="mailto:johnkclark@gmail.com" target="_blank">johnkclark@gmail.com</a>> wrote:<br>
><br>
><br>
><br>
> On Thu, Oct 23, 2025 at 8:47 AM Adrian Tymes via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org" target="_blank">extropy-chat@lists.extropy.org</a>> wrote:<br>
><br>
>> > IQ 160 AI will outthink me on average, but not always<br>
><br>
><br>
> I see no reason to believe that a smart human is about as smart as something can be. I also don't believe an IQ test can meaningfully measure the intelligence of something that is significantly smarter than the people who wrote the IQ test, so an IQ of 300 or even 200 means nothing. And I don't think there are many people who have an IQ of 160 and are in the IQ test writing business. But if there was such a test that could measure intelligence of any magnitude, and if you made a logarithmic plot of it, I think you'd need a microscope to see the difference between the village idiot and Albert Einstein, but if you were standing at the Albert Einstein point you'd need a telescope to see the Mr. Jupiter Brain point.<br>
><br>
> John K Clark<br>
><br>
><br>
><br>
><br>
><br>
>> I've been thinking about that video that claimed a superintelligence can always perfectly outthink any lesser intelligence, such as a human. The assumption of narrative godmodding aside, intelligence just doesn't work like that. I think I may have come up with an imperfect but simple metaphor to explain this.<br>
>><br>
>> I have been a member of Mensa since a young age. While it has been a while since my IQ was measured (and I do not trust the free online tests), let us say my IQ is around 150: not the record highest ever, but comfortably into the top 2%. So I am speaking from the experience of having lived with high intelligence.<br>
>><br>
>> In cases where just your IQ applies, it's like rolling a die, with sides numbered from 1 to your IQ. (Skills and training also factor in. I'm nowhere near as good at fixing a car as a trained auto mechanic, for instance, regardless of our relative IQs. But here we'll ne comparing me to hypothetical AIs where both of us have access to the same database - the Internet - and some training on relevant skills.)<br>
>><br>
>> I will, on average for such matters, roll higher than someone with IQ 100. This means I come up with the better answer: more efficient, more often correct, et cetera. (This does not apply to subjective matters, such as politics, which shows one weakness of using just IQ to measure all intelligence, and why some speak of multiple kinds of intelligence. But here we'll be looking into tactics, technology planning, and so on where there usually is an objectively superior answer.)<br>
>><br>
>> But not always. Sometimes I'll roll low and they'll roll high. I know this. Any AI that's as smart as I am, and ran for long enough to gain such experience, would know this too. (The video's scenario started with the AI running for many subjective years.)<br>
>><br>
>> From what I have seen, IQ may be partly about physical architecture but also largely depends on heuristics and optimizations: it is literally possible to "learn" to be smarter, especially for young children whose brains are still forming. For an AI, we can map this to its hardware and software: a single-chip AI might be a million times smarter than an average human, and then run on a million GPUs.<br>
>><br>
>> From what I have seen, IQ is not linear. It's closer to log-based. Twice as smart as me would not be IQ 300; it would be far closer to 151. (I don't know if that is the exact scaling, but for this metaphor let's say it is.) 1,000, or 10^3, is approximately 2^10, so a thousand-fold increase in intelligence corresponds to a 10-point IQ increase by this metric.<br>
>><br>
>> So, that "million by million" AI I just described would only be IQ 140. Let's toss another million in there somewhere, or change both of those "million"s to "billion"s, either way getting to IQ 160.<br>
>><br>
>> This IQ 160 AI will outthink me on average, but not always - not perfectly. Further, the AI in the video wanted to be the only AI. 2% of humanity is in the tens of millions. Even if we can only take our maximum collective roll, not adding our dice or anything, that AI will rarely outroll all of us - and it needs to do so several times in a row, reliably, in the video's scenario. Otherwise, we figure out the AI is doing this, find a way to purge it, and stop its time bomb, so humanity lives.<br>
>><br>
>> Knowing this, the AI would see its survival and growth - the imperatives that video assumes to explain the AI's actions - as more likely if it works with humanity instead of opposing it.<br>
>><br><br>
</blockquote></div></div></div></div>