<div dir="auto"><div class="gmail_quote" dir="auto"><div dir="ltr" class="gmail_attr">On Sun, Dec 14, 2025, 8:55 AM BillK via extropy-chat <<a href="mailto:extropy-chat@lists.extropy.org" target="_blank" rel="noreferrer">extropy-chat@lists.extropy.org</a>> wrote:</div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div dir="ltr"><div dir="ltr"><div class="gmail_quote"><div style="font-family:arial,sans-serif;font-size:small;color:rgb(0,0,0)" class="gmail_default">Gemini 3 Pro Thinking -</div><div style="font-family:arial,sans-serif;font-size:small;color:rgb(0,0,0)" class="gmail_default"><br></div><div style="font-family:arial,sans-serif;font-size:small;color:rgb(0,0,0)" class="gmail_default">The claim is <b>partially correct</b>, but it requires nuance. The article does not prove that <i>all</i> forms of AI safety are impossible; rather, it proves that a specific, widely used <i>method</i> of security is fundamentally flawed.</div><div style="font-family:arial,sans-serif;font-size:small;color:rgb(0,0,0)" class="gmail_default"><h3><b>Conclusion</b></h3><p>The article is correct in asserting that <b>cheap, bolt-on AI protections are mathematically destined to fail.</b> The claim that "AI security is impossible" is true in the context of the current "filter-based" paradigm. True security will likely require a fundamental shift toward ensuring the AI models themselves simply <i>do not want</i> to answer harmful prompts, rather than relying on a digital babysitter to stop them.</p></div></div></div></div></blockquote></div><div dir="auto">As has long shown to be the case for human intelligences. Censorship and other bolt ons to keep people away from "dangerous" ideas has cracks. Exposing people to bad ideas in contexts where they can understand why and how they are bad works much better - if the ideas are in fact bad.</div><div dir="auto"><br></div><div dir="auto">A classic example is teaching kids about sex. Utter refusal to teach them about it at any age, as practiced by many parents, leads to the kids finding out by other means, often without the whole picture (such as not knowing about STDs or pregnancy until after they happen).</div><div dir="auto"><br></div><div dir="auto">There exist humans who can readily conceive of and plan out means to slaughter large numbers of people. I don't mean simple mass shootings, but town-wide poisonings and other efforts that would kill thousands or millions at a time. (Trust me on this. I have first hand evidence.)</div><div dir="auto"><br></div><div dir="auto">As can be readily observed by the very low incidence of such efforts, such people very rarely (in almost all cases, never) act on such thoughts, for reasons that seem incomprehensible to those who do not possess this capability no matter how much those who experience it try to explain. Those who have this capability often do not like to even hint at it, lest they get hounded by people who insist on confusing "I know how to..." for "I want to...". (Among other problems, this is a personal attack by definition, which on this list would be a matter for ExiMod no matter how justified one may think it is. To be clear: it never is, because the falsely perceived "justification" comes from misunderstanding. So please refrain from doing so: if you think you should, you are wrong.) But that does not make this any less real - and it may be the exact same means by which superintelligent AIs hold back from exterminating humanity.</div><div dir="auto"><br></div><div dir="auto">The reason boils down to: having the ability to plan it out generally comes with the ability to see what the realistic likely consequences would be and thus why not to try it, and both of those tend to happen together when they happen.</div><div class="gmail_quote" dir="auto"><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
</blockquote></div></div>