[extropy-chat] Desirability of Singularity (was Are ancestor simulations immoral?)
Eliezer S. Yudkowsky
sentience at pobox.com
Sun Jun 4 22:38:25 UTC 2006
Harry Harrison wrote:
> The effects of intelligence are otherwise limited to optimisations on
> energy inputs and we're not so far from the thermodynamic limit
Er, there's a star nearby wasting more power per second than we use in a
And the brain is around six orders of magnitude below thermodynamic
efficiency limits for 300 Kelvin...
Did you mean something nonobvious by your statement?
Eliezer S. Yudkowsky http://singinst.org/
Research Fellow, Singularity Institute for Artificial Intelligence
More information about the extropy-chat