[ExI] Limiting factors of intelligence explosion speeds

Eugen Leitl eugen at leitl.org
Fri Jan 21 11:36:36 UTC 2011


On Thu, Jan 20, 2011 at 05:04:01PM -0500, Richard Loosemore wrote:

>> Unnecessary for darwinian systems. The process is dumb as dirt,
>> but it's working quite well.
>
> By even uttering the phrase "darwinian systems" you introduce a raft of  
> assumptions that all have huge implications.

Duh.

> This implies a process in which the entire AGI is improving as a result  
> of it being introduced into a big ecosystem of other AGIs, all exposed  

Yes, exactly.

> to a shaping environment, and with test-and-improve mechanisms akin to  
> those that operate in nature, to evolve biological systems.

And in memetic evolution, and in the economy.

> That is a GIGANTIC set of assumptions.  At the very least the AGIs have  

Yet it is the default mode of reality.

> to be made in very large numbers, to get meaningful competition and  
> fitness pressures ..... so if each is initially the size of a  

Speaking about assumptions.

> supercomputer, this process of improvement won't even start until we  
> have the resources to build thousands or millions of AGI supercomputers!

Think the Internet 30+ years hence.

> And, also, the fitness function has to do..... what, exactly?  Make them  

What, exectly, is your fitness function you're subject to? Can you
write a formal expression for it? Or do you deny you're even
subject to selection?

> compete for mates, according to the size of their computational muscles?  
>  Obviously a stupid idea, but if not that, then what?  And when do they  

Speaking about stupid ideas, you're pretty good at these.

> get killed off, in this evolutionary competition?  Are they being bred  
> for their logical abilities, their compassion, their sensitivity...?
>
> And assuming that they do not start with fantastically  
> greater-than-human thought speed, does this AGI evolutionary process  
> require them to live for an entire human maturation period before they  
> have babies?
>
> None of this makes any sense.

Try ideas that make sense, then.

> Somehow, the idea that genetic algorithms can improve their fitness has  
> exploded out of control and become the idea that AGIs can improve by an  
> evolutionary mechanism, but without any meaningful answers to these  
> critical questions about the evolutionary process ON THAT LARGER SCALE.
>
> What works for a GA -- the millisecond generation turnaround you mention  
> below -- is completely nuts for a full-blown AGI.

A million ms is about quarter an hour. A billion ms is two weeks.

> Let me state this clearly.  If an AGI can engage in meaningful  
> interaction with the world, sufficient for something to evaluate its  

Co-evolution is mostly others, embedded in an environment.
Rendering artificial reality at equivalent speedup is cheap.

> performance and decide on its "fitness", in the space of one  
> millisecond, then you already have a superintelligent AGI!  It would  
> probably already be operating at a million times human speed (and having  
> some pretty weird interactions with the universe, at that speed), if we  
> presume it to be capable of having its fitness judged after one 
> millisecond.
>
> I hereby declare this whole "build an AGI by darwinian evolution" idea  
> to be so logically incoherent that it does not need to detain us any 
> longer.

See, you're doing it again. But since you're a product of darwinian 
evolution, you're so logically incoherent that we can disregard whatever
you say. How convenient!

>
>
>
>
> Richard Loosemore
>
>
>>> usual.  I do not believe it makes sense to talk about what happens  
>>
>> If you define the fitness function, and have ~ms generation
>> turaround it's not quite as usual anymore.
>>
>>> *before* that point as part of the "intelligence explosion".
>>>
>>> D)  When such a self-understanding system is built, it is unlikely that
>>
>> I don't think that a self-understanding system is at all possible.
>> Or, rather, it would perform better than a blind optimization.
>>
>>> it will be the creation of a lone inventor who does it in their shed at
>>> the bottom of the garden, without telling anyone.  Very few of the "lone
>>> inventor" scenarios (the Bruce Wayne scenarios) are plausible.
>>
>> I agree it's probably a large scale effort, initially.
>>
>>> E)  Most importantly, the invention of a human-level, self-understanding
>>
>> I wonder where the self-understanding meme is coming from. It's  
>> certainly pervasive enough.
>>
>>> AGI would not lead to a *subsequent* period (we can call it the
>>> "explosion period") in which the invention just sits on a shelf with
>>> nobody bothering to pick it up.  A situation in which it is just one
>>> quiet invention alongside thousands of others, unrecognized and not
>>> generally believed.
>>>
>>> F)  When the first human-level AGI is developed, it will either require
>>> a supercomputer-level of hardware resources, or it will be achievable
>>
>> Bootstrap takes many orders of mangnitude more resources than required
>> for operation. Even before optimization happens.
>>
>>> with much less.  This is significant, because world-class supercomputer
>>> hardware is not something that can quickly be duplicated on a large
>>> scale.  We could make perhaps hundreds of such machines, with a massive
>>
>> About 30 years from now, TBit/s photonic networking is the norm.
>> The separation between core and edge is gone, and inshallah, so
>> well policy enforcement. Every city block is a supercomputer, then.
>>
>>> effort, but probably not a million of them in a couple of years.
>>
>> There are a lot of very large datacenters with excellent network
>> cross-section, even if you disregard large screen TVs and game  
>> consoles on >GBit/s residential networks.
>>
>>> G)  There are two types of intelligence speedup:  one due to faster
>>> operation of an intelligent system (clock speed) and one due to
>>
>> Clocks don't scale, eventually you'll settle for local asynchronous,
>> with large-scale loosely coupled oscillators synchronizing.
>>
>>> improvment in the type of mechanisms that implement the thought
>>> processes.  Obviously both could occur at once, but the latter is far
>>
>> How much random biochemistry tweaking would improve dramatically
>> on the current CNS performance? As a good guess, none. So once you've
>> reimplemented the near-optimal substrate, dramatic improvements
>> are over. This isn't software, this is direct implmenentation of
>> neural computational substrate in as thin hardware layer this
>> universe allows us.
>>
>>> more difficult to achieve, and may be subject to fundamental limits that
>>> we do not understand.  Speeding up the hardware, on the other hand, has
>>
>> I disagree, the limits are that of computational physics, and these are
>> fundamentally simple.
>>
>>> been going on for a long time and is more mundane and reliable.  Notice
>>> that both routes lead to greater "intelligence", because even a human
>>> level of thinking and creativity would be more effective if it were
>>> happening (say) a thousand times faster than it does now.
>>
>> Run a dog for a gigayear, still no general relativity.
>>
>>> *********************************************
>>>
>>> Now the specific factors you list.
>>>
>>> 1) Economic growth rate
>>>
>>> One consequence of the above reasoning is that economic growth rate
>>> would be irrelevant.  If an AGI were that smart, it would already be
>>
>> Any technology allowing you to keep a mind in a box will allow you
>> to make a pretty good general assembler. The limits of such technology
>> are energy and matter fluxes. Buying and shipping widgets is only a
>> constraining factor in the physical layer bootstrap (if at all necessary,
>> 30 years hence all-purpose fabrication has a pretty small footprint).
>>
>>> obvious to many that this was a critically important technology, and no
>>> effort would be spared to improve the AGI "before the other side 
>>> does".  Entire national economies would be sublimated to the goal of 
>>> developing  the first superintelligent machine.
>>
>> This would be fun to watch.
>>
>>> In fact, economic growth rate would be *defined* by the intelligence
>>> explosion projects taking place around the world.
>>>
>>>
>>> 2) Investment availability
>>>
>>> The above reasoning also applies to this case.  Investment would be   
>>> irrelevant because the players would either be governments or 
>>> frenzied
>>> bubble-investors, and they would be pumping it in as fast as money 
>>> could  be printed.
>>>
>>>
>>> 3) Gathering of empirical information (experimentation, interacting with
>>> an environment).
>>>
>>> So, this is about the fact that the AGI would need to do some   
>>> experimentation and interaction with the environment.  For example, 
>>> if  
>>
>> If you have enough crunch to run a mind, you have enough crunch to
>> run really really really good really fast models of the universe.
>>
>>> it wanted to reimplement itself on faster hardware (the quickest 
>>> route  to an intelligence increase) it would probably have to set up 
>>> its own  hardware research laboratory and gather new scientific data 
>>> by doing  experiments, some of which would go at their own speed.
>>
>> You're thinking like a human.
>>
>>> The question is:  how much of the research can be sped up by throwing
>>> large amounts of intelligence at it?  This is the parallel-vs-serial
>>> problem (i.e. you can't make a baby nine times quicker by asking nine 
>>>  women to be pregnant for one month).
>>
>> It's a good question. I have a hunch (no proof, nothing) that the
>> current way of doing reality modelling is extremely inefficient.
>> Currently, experimenters have every reason to sneer at modelers.
>> Currently.
>>
>>> This is not a factor that I believe we can understand very well ahead of
>>> time, because some experiments that look as though they require
>>> fundamentally slow physical processes -- like waiting for a silicon
>>> crystal to grow, so we can study a chip fabrication mechanism -- may
>>> actually be dependent on smartness, in ways that we cannot anticipate.
>>> It could be that instead of waiting for the chips to grow at their own
>>> speed, the AGI can do clever micro-experiments that give the same
>>> information faster.
>>
>> Any intelligence worth its salt would see that it would use computational
>> chemistry to bootstrap molecular manufacturing. The grapes could be hanging
>> pretty low there.
>>
>>> This factor invites unbridled speculation and opinion, to such an extent
>>> that there are more opinions than facts.  However, we can make one
>>> observation that cuts through the arguments.  Of all the factors that 
>>>  determine how fast empirical scientific research can be carried out, 
>>> we  know that intelligence and thinking speed of the scientist 
>>> themselves  *must* be one of the most important, today.  It seems 
>>> likely that in our  present state of technological sophistication, 
>>> advanced research  projects are limited by the availability and cost 
>>> of intelligent and  experienced scientists.
>>
>> You can also vastly speed up the rate of prototyping by scaling down
>> and proper tooling. You see first hints of that in lab automation,
>> particularly microfluidics. Add ability to fork off dedicated investigators
>> at the drop of a hat, and things start happening, and in a positive-feedback
>> loop.
>>
>>> But if research labs around the world have stopped throwing *more*
>>> scientists at problems they want to solve, because the latter cannot be
>>> had, or are too expensive, would it be likely that the same research
>>> labs ar *also*, quite independently, at the limit for the physical rate
>>> at which experiments can be carried out?  It seems very unlikely that 
>>>  both of these limits have been reached at the same time, because 
>>> they  cannot be independently maximized.  (This is consistent with 
>>> anecdotal  reports:  companies complain that research staff cost a 
>>> lot, and that  scientists are in short supply:  they don't complain 
>>> that nature is just  too slow).
>>
>> Most monkeys rarely complain that they're monkeys. (Resident monkeys
>> excluded, of course).
>>
>>> In that case, we should expect that any experiment-speed limits lie up
>>> the road, out of sight.  We have not reached them yet.
>>
>> I, a mere monkey, can easily imagine two orders of magnitude speed
>> improvements. Which, of course, result in a positive autofeedback loop.
>>
>>> So, for that reason, we cannot speculate about exactly where those
>>> limits are.  (And, to reiterate:  we are talking about the limits that
>>> hit us when we can no longer do an end-run around slow experiments by
>>
>> I do not think you will need slow experiments. Not slow by our standards,
>> at least.
>>
>>> using our wits to invent different, quicker experiments that give the
>>> same information).
>>>
>>> Overall, I think that we do not have concrete reasons to believe that
>>> this will be a fundamental limit that stops the intelligence explosion
>>> from taking an AGI from H to (say) 1,000 H.  Increases in speed within
>>> that range (for computer hardware, for example) are already expected,
>>> even without large numbers of AGI systems helping out, so it would 
>>> seem  to me that physical limits, by themselves, would not stop an 
>>> explosion  that went from I = H to I = 1,000 H.
>>
>> Speed limits (assuming classical computation) do not begin to take hold
>> before 10^6, and maybe even 10^9 (this is more difficult, and I do not
>> have a good model of wetware at 10^9 speedup to current wallclock).
>>
>>> 4)  Software complexity
>>>
>>> By this I assume you mean the complexity of the software that an AGI
>>> must develop in order to explode its intelligence.  The premise is
>>> that even an AGI with self-knowledge finds it hard to cope with the
>>> fabulous complexity of the problem of improving its own software.
>>
>> Software, that's pretty steampunk of you.
>>
>>> This seems implausible as a limiting factor, because the AGI could
>>> always leave the software alone and develop faster hardware.  So long as
>>
>> There is no difference between hardware and software (state) as far
>> as advanced cognition is concerned. Once you've covered the easy
>> gains in first giant co-evolution steps further increases are much
>> more modest, and much more expensive.
>>
>>> the AGI can find a substrate that gives it (say) 1,000 H thinking-speed,
>>
>> We should be able to do 10^3 with current technology.
>>
>>> we have the possibility for a significant intelligence explosion.
>>
>> Yeah, verily.
>>
>>> Arguing that software complexity will stop the initial human level 
>>> AGI  
>>
>> If it hurts, stop doing it.
>>
>>> from being built is a different matter.  It may stop an intelligence  
>>> explosion from happening by stopping the precursor events, but I take 
>>>  that to be a different type of question.
>>>
>>>
>>> 5)  Hardware demands vs. available hardvare
>>>
>>> I have already mentioned, above, that a lot depends on whether the 
>>> first  AGI requires a large (world-class) supercomputer, or whether 
>>> it can be  done on something much smaller.
>>
>> Current supercomputers are basically consumer devices or embeddeds on  
>> steroids, networked on a large scale.
>>
>>> This may limit the initial speed of the explosion, because one of the 
>>>  critical factors would be the sheer number of copies of the AGI that 
>>> can  
>>
>> Unless the next 30 years won't see the same development as the last ones,
>> then substrate is the least of your worries.
>>
>>> be created.  Why is this a critical factor?  Because the ability to 
>>> copy  the intelligence of a fully developed, experienced AGI is one 
>>> of the big  new factors that makes the intelligence explosion what it 
>>> is:  you  cannot do this for humans, so human geniuses have to be 
>>> rebuilt from  scratch every generation.
>>>
>>> So, the initial requirement that an AGI be a supercomputer would make 
>>> it  hard to replicate the AGI on a huge scale, because the 
>>> replication rate  would (mostly) determine the 
>>> intelligence-production rate.
>>
>> Nope.
>>
>>> However, as time went on, the rate of replication would grow, as  
>>
>> Look, even now we know what we would need, but you can't buy it. But  
>> you can design it, and two weeks from now you'll get your first 
>> prototypes.
>> That's today, 30 years the prototypes might be hours away.
>>
>> And do you need prototypes to produce a minor variation on a stock
>> design? Probably not.
>>
>>> hardware costs went down at their usual rate.  This would mean that 
>>> the  *rate* of arrival of high-grade intelligence would increase in 
>>> the years  following the start of this process.  That intelligence 
>>> would then be  used to improve the design of the AGIs (at the very 
>>> least, increasing  the rate of new-and-faster-hardware production), 
>>> which would have a  positive feedback effect on the intelligence 
>>> production rate.
>>>
>>> So I would see a large-hardware requirement for the first AGI as   
>>> something that would dampen the initial stages of the explosion.  But 
>>>  
>>
>> Au contraire, this planet is made from swiss cheese. Annex at your leisure.
>>
>>> the positive feedback after that would eventually lead to an 
>>> explosion  anyway.
>>>
>>> If, on the other hand, the initial hardware requirements are modest 
>>> (as  they very well could be), the explosion would come out of the 
>>> gate at  full speed.
>>>
>>>
>>>
>>>
>>> 6)  Bandwidth
>>>
>>> Alongside the aforementioned replication of adult AGIs, which would   
>>> allow the multiplication of knowledge in ways not currently available 
>>> in  humans, there is also the fact that AGIs could communicate with 
>>> one  another using high-bandwidth channels.  This inter-AGI 
>>> bandwidth.
>>
>> Fiber is cheap. Current fiber comes in 40 or 100 GBit/s parcels.
>> 30 years hence bandwidth will be probably adequate.
>>
>>> As a separate issue, there might be bandwidth limits inside an AGI,   
>>> which might make it difficult to augment the intelligence of a single 
>>>  system.  This is intra-AGI bandwidth.
>>
>> Even now bandwidth growth is far in excess of computation growth.
>> Once you go embedded memory, you're more closely matched. But still
>> the volume/surface (you only have to communicate surface state)
>> ratio indicated the local communication is the bottleneck.
>>
>>> The first one - inter-AGI bandwidth - is probably less of an issue 
>>> for  the intelligence explosion, because there are so many research 
>>> issues  that can be split into separably-addressible components, that 
>>> I doubt we  would find AGIs sitting around with no work to do on the 
>>> intelligence  amplification project, on account of waiting for other 
>>> AGIs to get a  free channel to talk to them.
>>
>> You're making it sound so planned, and orderly.
>>
>>> Intra-AGI bandwidth is another matter entirely.  There could be   
>>> limitations on the IQ of an AGI -- for example if working memory   
>>> limitations (the magic number seven, plus or minus two) turned out to 
>>> be  caused by connectivity/bandwidth limits within the system.
>>
>> So many assumptions.
>>
>>> However, notice that such factors may not inhibit the initial phase 
>>> of  an explosion, because the clock speed, not IQ, of the AGI may be  
>>> 
>>
>> There is no clock, literally. Operations/volume, certainly.
>>
>>> improvable by several orders of magnitude before bandwidth limits 
>>> kick  in.  The reasoning behind this is the observation that neural 
>>> signal  
>>
>> Volume/surface ratio is on your side here.
>>
>>> speed is so slow.  If a brain-like system (not necessarily a whole 
>>> brain  emulation, but just something that replicated the high-level   
>>> functionality) could be built using components that kept the same 
>>> type  of processing demands, and the same signal speed.  In that kind 
>>> of  system there would then be plenty of room to develop faster 
>>> signal  speeds and increase the intelligence of the system.
>>>
>>> Overall, this is, I believe, the factor that is most likely to cause  
>>> trouble.  However, much research is needed before much can be said 
>>> with  certainty.
>>>
>>> Most importantly, this depends on *exactly* what type of AGI is being 
>>>  built.  Making naive assumptions about the design can lead to false  
>>> conclusions.
>>
>> Just think of it as a realtime simulation of a given 3d physical
>> process (higher dimensions are mapped to 3d, so they don't figure).
>> Suddenly things are simple.
>>
>>>
>>> 7)  Lightspeed lags
>>>
>>> This is not much different than bandwidth limits, in terms of the 
>>> effect  it has.  It would be a significant problem if the components 
>>> of the  machine were physically so far apart that massive amounts of 
>>> data (by  assumption) were delivered with a significant delay.
>>
>> Vacuum or glass is a FIFO, and you don't have to wait for ACKs.
>> Just fire stuff bidirectionally, and deal with transmission errors
>> by graceful degradation.
>>
>>> By itself, again, this seems unlikely to be a problem in the initial 
>>> few  orders of magnitude of the explosion.  Again, the argument 
>>> derives from  what we know about the brain.  We know that the brain's 
>>> hardware was  chosen due to biochemical constraints.  We are 
>>> carbon-based, not  silicon-and-copper-based, so, no chips in the 
>>> head, only pipes filled  with fluid and slow molecular gates in the 
>>> walls of the pipes.  But if  nature used the pipes-and-ion-channels 
>>> approach, there seems to be  plenty of scope for speedup with a 
>>> transition to silicon and copper (and  never mind all the other more 
>>> exotic computing substrates on the  horizon).  If that transition 
>>> produced a 1,000x speedup, this would be  an explosion worthy of the 
>>> name.
>>
>> Why so modest?
>>
>>> The only reason this might not happen would be if, for some reason, 
>>> the  brain is limited on two fronts simultaneously:  both by the 
>>> carbon  implementation and by the fact that bigger brains cause 
>>> disruptive  
>>
>> The brain is a slow, noisy (but one using noise to its own advantage)
>> metabolically constrained system which burns most of its metabolism  
>> for homeostasis purposes. It doesn't take a genius to sketch the
>> obvious ways in which you can reimplement that design, taking advantages
>> and removing disadvantages.
>>
>>> light-speed delays.  Or, that all non-carbon-implementation of the 
>>> brain  take us up close to the lightspeed limit before we get much of 
>>> a speedup  
>>
>> We here work with ~120 m/s, not 120 Mm/s. Reduce feature size by
>> an order of magnitude or two, and switching times of ns and ps
>> instead of ms, and c is not that big a limitation anymore.
>>
>>> over the brain.  Neither of these ideas seem plausible.  In fact, 
>>> they  both seem to me to require a coincidence of limiting factors 
>>> (two  limiting factors just happening to kick in at exactly the same 
>>> level),  which I find deeply implausible.
>>>
>>>
>>> *****************
>>>
>>> Finally, some comments about approaches to AGI that would affect the  
>>> answer to this question about the limiting factors for an 
>>> intelligence  explosion.
>>>
>>> I have argued consistently, over the last several years, that AI   
>>> research has boxed itself into a corner due to a philosophical   
>>> commitment to the power of formal systems.  Since I first started  
>>
>> Very much so.
>>
>>> arguing this case, Nassim Nicholas Taleb (The Black Swan) coined the  
>>> term "Ludic Fallacy" to describe a general form of exactly the issue 
>>> I  have been describing.
>>>
>>> I have framed this in the context of something that I called the   
>>> "complex systems problem", the details of which are not important 
>>> here,  although the conclusion is highly relevant.
>>>
>>> If the complex systems problem is real, then there is a very large 
>>> class  of AGI system designs that are (a) almost completely ignored 
>>> at the  moment, and (b) very likely to contain true intelligent 
>>> systems, and (c)  quite possibly implementable on relatively modest 
>>> hardware.  This class  
>>
>> Define "relatively modest".
>>
>>> of systems is being ignored for sociology-of-science reasons (the   
>>> current generation of AI researchers would have to abandon their 
>>> deepest  loves to be able to embrace such systems, and since they are 
>>> fallible  humans, rather than objectively perfect scientists, this is 
>>> anathema).
>>
>> Which is why blind optimization processes running on acres of
>> hardware will kick their furry little butts.
>>
>>> So, my most general answer to this question about the rate of the   
>>> intelligence explosion is that, in fact, it depends crucially on the  
>>> kind of AGI systems being considered.  If the scope is restricted to 
>>> the  current approaches, we might never actually reach human level   
>>> intelligence, and the questio is moot.
>>>
>>> But if this other class of (complex) AGI systems did start being 
>>> built,  we might find that the hardware requirements were relatively 
>>> modest  (much less than supercomputer size), and the software 
>>> complexity would  also not be that great.  As far as I can see, most 
>>> of the  
>>
>> I love this "software" thing.
>>
>>> above-mentioned limitations would not be significant within the first 
>>>  few orders of magnitude of increase.  And, the beginning of the 
>>> slope  could be in the relatively near future, rather than decades 
>>> away.
>>
>> In order to have progress, you first have to have people working on it.
>>
>>> But that, as usual, is just the opinion of an AGI researcher.  No 
>>> need  to take *that* into account in assessing the factors.  ;-)
>>
>> Speaking of AGI researchers: do you have a nice publication track of
>> yours you could dump here? 
>>
>
> _______________________________________________
> extropy-chat mailing list
> extropy-chat at lists.extropy.org
> http://lists.extropy.org/mailman/listinfo.cgi/extropy-chat
-- 
Eugen* Leitl <a href="http://leitl.org">leitl</a> http://leitl.org
______________________________________________________________
ICBM: 48.07100, 11.36820 http://www.ativel.com http://postbiota.org
8B29F6BE: 099D 78BA 2FD3 B014 B08A  7779 75B0 2443 8B29 F6BE



More information about the extropy-chat mailing list