Hi Matt,

First, you should have gotten an A on your paper for such progressive ideas
at the time.

Second, if you ask these questions, this means that I have completely
failed to explain my proposal. Had I managed to explain the idea, then you
would have not asked those questions.

So, Hebb's rule is about plasticity or in other words, about learning. What
I am proposing is not about plasticity or learning. This is where the
difference lies. The gates open and close connections but only transiently
for example, only for one second and then they go back to their default
state. This is something different from synapses and in addition to
synapses/weights.

This is a completely new type of a mechanism, or a new level of complexity
that did not exist before.

I hope this helps.

Danko


Dr. Danko Nikolić
www.danko-nikolic.com
https://www.linkedin.com/in/danko-nikolic/
-- I wonder, how is the brain able to generate insight? --


On Wed, Oct 18, 2023 at 5:22 PM Matt Mahoney <mattmahone...@gmail.com>
wrote:

> How is your proposal different from Hebb's rule? I remember reading in the
> 1970s as a teenager about how neurons represent mental concepts and
> activate or inhibit each other through 2 kinds of synapses. I had the idea
> that synapses would change states in the process of forming memories. At
> the time it was unknown that synapses would do this. In 1980 I described
> this model of classical conditioning in my freshman psychology class. I got
> a B on the paper. Years later I learned that Hebb proposed the same idea in
> 1949.
>
> Connectionism is a simple idea that makes it easy to understand how brains
> work by learning associations between concepts, but it lacks a mechanism
> for adding new concepts. That problem is solved by representing concepts as
> linear combinations of neurons, but it makes a neural network more like a
> black box of inscrutable matrices.
>
> Your diagram shows a feed forward network, but in reality there are
> connections going in all directions. Lateral inhibition within the same
> layer gives you a winner take all network as the mechanism for attention in
> a transformer. Positive feedback loops give you short term memory. Negative
> feedback gives you logarithmic scaling of sensory perceptions.
>
> On Wed, Oct 18, 2023, 10:14 AM Danko Nikolic <danko.niko...@gmail.com>
> wrote:
>
>> Here is my proposal on what connectionism is missing in order to reach a
>> 'true" AI i.e., an AI that is much more similar to how the human brain
>> works.
>>
>> This is a nine-minute video on the secret that was missing:
>> https://www.youtube.com/watch?v=GVW6H4iCsTg&ab_channel=dankonikolic
>>
>> I hope it is clear enough.
>>
>> Comments and questions are welcome.
>>
>> Danko
>>
>>
>> Dr. Danko Nikolić
>> www.danko-nikolic.com
>> https://www.linkedin.com/in/danko-nikolic/
>> -- I wonder, how is the brain able to generate insight? --
>>
> *Artificial General Intelligence List <https://agi.topicbox.com/latest>*
> / AGI / see discussions <https://agi.topicbox.com/groups/agi> +
> participants <https://agi.topicbox.com/groups/agi/members> +
> delivery options <https://agi.topicbox.com/groups/agi/subscription>
> Permalink
> <https://agi.topicbox.com/groups/agi/T1e8c992894d0432b-M35928e49c51d76b5e6625ed3>
>

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/T1e8c992894d0432b-M129e36835dc9e0f8ab6b4b1f
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to