[ExI] Mind Uploading: is it still me?
Ben Zaiboc
ben at zaiboc.net
Sat Dec 27 23:05:32 UTC 2025
On 26/12/2025 12:59, John Clark wrote:
> In 2024 researchers at Harvard and Google sliced one cubic millimeter
> of a human brain into 5019 slices and then used an electron microscope
> to take photographs of each slice, they then used the resulting 1.4
> petabytes (1,400 trillion bytes) of data to construct a 3-D model of
> that cubic millimeter. They got a X-Y resolution of 4 nanometers (a
> DNA molecule is 2.5 nanometers thick) but the Z resolution(thickness)
> was only 30 nanometers. That probably would be good enough
> resolutionfor an upload but unfortunately it was just for one cubic
> millimeter, the average human brain contains about 1,400,000 cubic
> millimeters.
On the surface, this sounds quite discouraging, but I'd just like to
make a brain-dump on the topic, having been thinking about it for a while:
The x-y resolution mentioned was overkill, by at least 10 times, and the
z resolution less so, but still probably higher than necessary. Let's
say it was just about right, though. That means approx. 14 trillion
bytes for 1 cubic mm.
Times that by 1.4M for the whole brain (almost certainly not actually
needed (for our purposes), for several reasons, and as discussed
previously, a generic human brain model could probably cut down on that
considerably, with individual variations laid on top of it), so we get
14 x 10^12 times 1.4 x 10^6 = 19.6x10^18 bytes (? please check this, I'm
on dodgy ground here, given my mathematical challenges). Say around 20
exabytes.
That's a lot, but I reckon it can be reduced a fair bit (a lot,
actually), considering that most of the brain is white matter (axons),
not grey matter (cortex, where the vast majority of the synaptic
information needs to come from), and all that's needed for the white
matter is to record each axon's origin and destination (including
branchings), and axons tend to be about 300nm in diameter, so the
resolution doesn't have to be so high as for the cortex. Basically,
what's needed for the bulk of the brain is a wiring diagram (a whopping
big, complex one, but still nothing more than that. Probably).
Plus the fact that the original resolution of the scan wouldn't be the
same as the resolution of the data derived from it, and therefore the
storage needed. Lossless compression algorithms as well as dynamic
intelligent feature extraction would reduce the amount of scanned data
on the fly, so you might be scanning 14 trillion bytes per cubic mm but
condensing that into a few megabytes or less, without losing anything
relevant to reconstructing the neural structure and function. There'd be
no point recording every 300nm voxel of white matter to produce a
connectome map because that would be like creating a raster diagram when
a vector diagram is far more efficient. I think there are parallels here
with the way our brains process vision, extracting and compactly
representing specific features of our visual fields, and ignoring the rest.
Yes, the scanning would have to be very detailed, but as soon as a
structure can be detected, and linked to prevous structure, that
detailed data can be discarded.
When you think that we'd be effectively scanning trillions of organic
molecules, and converting their positions into a datastream, you can see
there's plenty of scope for compression, just for starters.
Unknowns include whether variations in the myelin sheaths are important,
and whether the support cells (glial cells) need to be recorded as well
or not. My money's on Not.
It would probably be very useful to devise a scheme where different
aspects of the brain scan are condensed into separate maps, that can
each be stored in a relatively compact way, to later be integrated and
expanded into a functioning structure when it comes time to actually
build the upload.
So overall, I expect that the 20 exabytes estimate to capture all the
relevant data from an individual brain is vastly greater than what will
eventually be really needed, once we figure out what needs to be
captured, how to encode and store it, and what can be recorded as
variations on a 'canonical human brain' model (which could potentially
make /enormous/ savings on the amount of data needed to encode an
individual brain (if that turns out to be a feasible idea!)).
Another factor will be the time needed to scan an entire brain. And
there's also the problem of the scanning method dumping heat into the
tissue surrounding the area being scanned, potentially messing up the
structure and chemical environment.
We probably need entirely different scanning methods to what we
currently have, or it wouldn't be remotely practical. I've no idea how
to approach that problem.
All this makes me more convinced than ever that cryonic suspension alone
won't be enough to prepare a brain for the kind of scanning that will be
needed for uploading. Aldehyde stabilisation, or something equivalent,
might just be the start.
Now I'm wondering if and how a brain could be dismantled - as in
separated into distinct sections - without damaging anything essential,
either before or after preservation. Then each section could be
separately scanned, in parallel. Hmmm...
If axons were elastic, I wonder if a 'reverse convolution' could be done
on the cortex. Cortices, rather. I expect cerebellum would need to be
uploaded as well, which presents its own problems (smaller neurons) and
opportunities (much more regular structure than elsewhere in the brain).
Or, if the axons could somehow have markers set in their connections to
the cortical layers, then be removed for separate scanning, then the
cortices spread out into what, four or five big thin sheets ??
Any other ideas?
--
Ben
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20251227/39693db6/attachment.htm>
More information about the extropy-chat
mailing list