[ExI] another open letter, but this one may be smarter than the previous one

Stuart LaForge avant at sollegro.com
Sun Apr 30 19:31:02 UTC 2023

Quoting Will Steinberg via extropy-chat <extropy-chat at lists.extropy.org>:

> Yeah, I kind of doubt we will make any meaningful progress on either of
> those descriptions of consciousness before AGI exists.  (And I *like* woo
> shit) Phenomena are completely inscrutable within the modern physics
> framework, and even for the "entity that can model itself" form, what does
> it even mean for a computer program to model itself like humans model
> ourselves?  It has no place, no body.  We don't even understand what is
> inside these LLMs in the first place...

Sure we do. A few million 300-dimensional vectors, representing words,  
organized into clusters in hyperspace to represent statistical  
relationships that act as proxy for meaning, or simply are meaning,  
depending on ones point of view. It is nature that is the blind  
watchmaker. We create with our eyes wide open.

> What a terrifying time to be alive.  I don't see a plausible scenario where
> this all doesn't lead to unbelievable amounts of suffering (of both
> biological and machine consciousness.)

In the best of times, some suffering is unavoidable, let alone in  
times of great change. But fear in anticipation of suffering is  
premature suffering begun needlessly. Be at peace, you have all the  
tools you need to survive this.


In the video, Dr. Li Jiang is a director of Stanford AIRE program  
gives some practical tips on how to survive the era of generative AI.

Stuart LaForge

More information about the extropy-chat mailing list