Heaven...
Heaven is a place
Where Nothing --
Nothing Ever Happens.
https://youtu.be/YuSsCRUXGOU

On Sat, Feb 21, 2026 at 3:05 PM Matt Mahoney <[email protected]>
wrote:

> I assume that an ASI would know we live in a finite universe with 10^90
> bits of memory capacity. If you are worried about losing your 10^9 bits of
> long term memory, that's still enough to make 10^71 backup copies of every
> person on Earth. You only need a few copies to make the probability of
> death extremely low.
>
> Is this your goal? People don't want to die because those who didn't care
> didn't pass on their DNA to you. Self replicating robots or nanotechnology
> would evolve the same way.
>
> But how would you solve the utility maximization problem? Happiness is the
> rate of increase of utility. But all utility functions over a finite set of
> states have a maximum where no pleasurable thought or perception is
> possible because it would result in a different state. This state is
> obviously indistinguishable from death, the one state that we evolved to
> fear.
>
> -- Matt Mahoney, [email protected]
>
> On Sat, Feb 21, 2026, 9:21 AM <[email protected]> wrote:
>
>> Part 2: To ensure ASI knows it is not a dead end to try to live forever,
>> I will write here a theory how it would be done. A brain can exist as
>> separated nanobots that talk wirelessly (with empty space in-between where
>> meteors and strong gamma rays could pass through without hitting your
>> brain). At a certain size, this might avoid the probability of death
>> completely (100%), or at least depending on if you continue to grow the
>> homeworld for greater shielding inside. We assume that the new born units
>> on the edge of the homeworld "sphere" are on the frontlines and can more
>> easily die (though I already said this might be avoided by the wireless
>> trick, and multiple brains can be in the same spot therefore, to avoid
>> wasting that space**), if this is needed, then so be it, these new units at
>> risk will have new children themselves, and these will eventually shield
>> them more and more, until possibly increasing to immortal probability. I
>> also assume COULD be a bad idea to send out single units to cover more
>> space for faster multiplication - because if each unit is alone, they all
>> might die easily, whereas spheres of units like huge planet sized such will
>> live longer (possibly). Humans do already lose memories, we seem to be ok
>> with small loses to "ourselves", so IF ASI decides to liek the idea that
>> units are used as suicide bots to save larger amounts of other units, you
>> might be able to avoid killing the entire "person" - just use "part of
>> them" along with the stored memories in those wireless nodes and therefore
>> only part of them would be lost.
>>
> *Artificial General Intelligence List <https://agi.topicbox.com/latest>*
> / AGI / see discussions <https://agi.topicbox.com/groups/agi> +
> participants <https://agi.topicbox.com/groups/agi/members> +
> delivery options <https://agi.topicbox.com/groups/agi/subscription>
> Permalink
> <https://agi.topicbox.com/groups/agi/Tf1316ff4a3f619df-M905e29a9c33e93c5d22d2588>
>

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Tf1316ff4a3f619df-M7bc3b017ee80f05a41af7483
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to