Jim,

>> "You keep confusing source with destination, because you insist on
>> operating within your declarative memory, which is a rather
>> superficial subset of your cognitive model :)."
>
> Are you replying using your theory as a model of the mind (indeed, as
> a model of my mind!) 

It's not *my* theory, a mainstream position in neuroscience is that neocortex 
is a hierarchy of generalization, from primary sensory & motor areas to 
incrementally higher association areas. It's also well known that declarative 
memory is restricted to the latter. Besides, these things are tautologically 
self-evident to me.

> with a smiley face to represent some humor about doing that?

That mostly represents my self-satisfaction with putting things well :).   

> And, are you saying that declarative memory is a destination in your
> model rather than a source? Is declarative memory derived?  That is
> what you are saying right?

Yes, see the above. If you want a mainstream source, read "Cortex & Mind" by 
Joaquin Fuster, he is a paramount authority on the subject.

> Is your theory a theory of how the brain works, a theory for
> artificial general intelligence using computers or both?

Both, but the artificial version is a whole lot cleaner, the brain is loaded 
with evolutionary artifacts. For example, I don't have this artificial 
distinction between implicit & declarative memory, between sensory & motor 
hierarchies, & a bunch of other things.

> Do you regularly see the kinds of thinking that people do in the terms
> of your model?

Yes, except that "my" part of it is well below the surface (low-level 
processing), the mainstream part is usually sufficient to qualitatively explain 
declarative thinking.

 http://www.cognitivealgorithm.info/2012/01/cognitive-algorithm.html  

--------------------------------------------------
From: "Jim Bromer" <[email protected]>
Sent: Wednesday, August 22, 2012 9:42 AM
To: "AGI" <[email protected]>
Subject: [agi] Boris Explains His Theory

> Boris,
> I am just not getting this.  So let me try starting with some simple 
> questions.
> I had said, "Forcing semantic values into 3-dimensional orthogonal
> space seems amazingly confused to me."
> You replied,
> "You keep confusing source with destination, because you insist on
> operating within your declarative memory, which is a rather
> superficial subset of your cognitive model :)."
> 
> Are you replying using your theory as a model of the mind (indeed, as
> a model of my mind!) with a smiley face to represent some humor about
> doing that?  Did you think that my statement about forcing semantic
> values was made in reference to something in your theory?  Because
> that is not what I meant.  I was just saying that I have read papers
> about using semantic vectors and my thoughts on that is that trying to
> force semantic vectors into 3-dimensional space seems confused.
> 
> And, are you saying that declarative memory is a destination in your
> model rather than a source? Is declarative memory derived?  That is
> what you are saying right?
> 
> Is your theory a theory of how the brain works, a theory for
> artificial general intelligence using computers or both?
> 
> Do you regularly see the kinds of thinking that people do in the terms
> of your model?
> Jim Bromer
> 
> 
> 
> --------------- Previous Messages ---------------
> Jim,
>> I don't understand your comments about detecting patterns. You said:
> 
> This is interactive pattern projection, but you have to discover those
> patterns first. Technically, you simply multiply all the vectors in a
> pattern by a relative distance to a target coordinate. And then you
> compare multiple patterns projected to the same coordinate, & multiply
> the difference by relative strength of each pattern. That gives you a
> combined prediction, or probability distribution if the patterns are
> mutually exclusive.
> 
> That comment was about projecting patterns, not detecting them.
> 
>> What kind of patterns are you talking about? How do the elemental 
>> observations (from the sensory device) get turned into vectors?
> 
> Comparisons generate derivatives. A vector is d(input) over
> d(coordinate). Conventionally, it's over multiple coordinates
> (dimensions), & the input can be a lower coordinate, but that's not
> essential.
> 
>> Are you saying that the "higher level of search and generalization" are 
>> where/how the pattern vectors are created?
> 
> No, all levels.
> 
>> Why or how would you pick out a particular target coordiate to use to 
>> combine a prediction?
> 
> Well, coordinate resolution is variable, so I am talking about a
> min->max span. Basically, vector projection is part of input selection
> for a higher-level search. The target coordinate span is a feedback
> from that higher level, or, if there aren't any, current_search_span *
> selection_rate: preset lossiness / sparseness of representation on the
> higher level.
> 
>> Are you saying that all predictions have individual coordinates?
> 
> Individual coordinate span. It's what + where, you got to have both.
> 
>> That alone means that they would have to exist in dynamic virtual space of 
>> many dimensions. Forcing semantic values into 3-dimensional orthogonal space 
>> seems amazingly confused to me.
> 
> You keep confusing source with destination, because you insist on
> operating within your declarative memory, which is a rather
> superficial subset of your cognitive model :).
> 
> We *derive* all our "semantic" values from 4D-continuous observation,
> no need to "force" them into it.
> 
>> What kind of space would your vectors exist in, how do they get there and 
>> why do you choose a particular coordinate for a combination of predictions?
> 
> As I said, hierarchical search generates incremental syntax, &
> variables within it are individually evaluated for search on
> successive levels. The strongest variable, whether it's an original
> coordinate | modality or a derivative thereof, becomes a coordinate
> for a higher level. The strength here must be averaged over higher
> level span.
> 
> It's hard to explain this on "semantic" level, which is profoundly
> confused in humans anyway. But a good intermediate example is Periodic
> Table. You take atomic mass (which is a derived, not an original
> variable) as top coordinate, compare pH value along that coordinate, &
> notice recurrent periodicity in it's variation. Since pH is a main
> chemical property, you then use it as a primary dimension that defines
> a period, & atomic mass becomes a secondary dimension that defines a
> sequence of periods. Both dimensions are derived, they may seem kind
> of a halfway between original & "semantic", but the same derivation
> process will get you to the latter
> 
> http://www.cognitivealgorithm.info/2012/01/cognitive-algorithm.html
> 
> Boris,
> 
> I don't understand your comments about detecting patterns. You said:
> 
> This is interactive pattern projection, but you have to discover those
> patterns first. Technically, you simply multiply all the vectors in a
> pattern by a relative distance to a target coordinate. And then you
> compare multiple patterns projected to the same coordinate, & multiply
> the difference by relative strength of each pattern. That gives you a
> combined prediction, or probability distribution if the patterns are
> mutually exclusive :).
> 
> What kind of patterns are you talking about?  How do the elemental
> observations (from the sensory device) get turned into vectors?  Are
> you saying that the "higher level of search and generalization" are
> where/how the pattern vectors are created? Why or how would you pick
> out a particular target coordiate to use to combine a prediction?  Are
> you saying that all predictions have individual coordinates?
> 
> I have read papers on Semantic Vectors, (I do not need to be told that
> the sources of semantic vectors are different than the sources of the
> products of your system) and I have always felt that they were
> absurdly inappropriate for semantics (or concepts) because they forced
> the semantic concepts into a system that they did not fit into.  As is
> so obvious to Two-Door, concepts are relativistic. That alone means
> that they would have to exist in dynamic virtual space of many
> dimensions.  Forcing semantic values into 3-dimensional orthogonal
> space seems amazingly confused to me.
> 
> What kind of space would your vectors exist in, how do they get there
> and why do you choose a particular coordinate for a combination of
> predictions?
> 
> (Incidentally, just to remind you, my ideas of concepts are not
> necessarily expressed as vectors although I am not close minded about
> the idea.)
> 
> Jim Bromer
> 
> 
> On Tue, Aug 21, 2012 at 2:22 PM, Boris Kazachenko <[email protected]> wrote:
> 
>> On the other hand I am interested in conjectures about conceptual vectors 
>> and stuff like that
> 
> You can't formalize "conceptual" vectors, except in terms of
> "conceptual" coordinates .
> 
> Jim Bromer
> 
> Thanks for the smiley faces Boris...
> I disagree that you have to   multiply all the vectors in a pattern by
> a relative distance to a target   coordinate in order to combine
> imagined complex ideas and related   observations. Our theories are
> very different. (On the other hand I am   interested in conjectures
> about conceptual vectors and stuff like that.)
> 
> I am interested in a continuation of the explanation of your theories
> and   I hope to get back to it soon.
> Jim Bromer
> 
> 
> On Tue, Aug 21, 2012 at 7:57 AM, Boris Kazachenko <[email protected]> wrote:
> 
> Jim,
> 
>>Where Boris and I disagree is that I feel that     because of relativity the 
>>input source of an idea may not be the most     elemental source of the idea 
>>that needs to be considered.
> 
> Right, but that's the simplest assumption, you must make     it unless
> you know otherwise. And you only know otherwise if you've
> discovered more "elemental" (stable) source on some higher level of
> search     & generalization. That would generate a focusing / motor
> feedback,     always derived from prior feedforward. As I keep saying,
> complexity must be     incremental :).
> 
>> One simple example is that we can use our     imagination and study of the 
>> subject of the concept in order to extend our     ideas about the subject 
>> beyond those ideas which came directly from     observations of it.
> 
> This is interactive pattern projection, but you have to     discover
> those patterns first. Technically, you simply multiply all the
> vectors in a pattern by a relative distance to a target coordinate.
> And then     you compare multiple patterns projected to the same
> coordinate, &     multiply the difference by relative strength of each
> pattern. That gives you     a combined prediction, or probability
> distribution if the patterns are     mutually exclusive :).
> 
> 
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/18407320-d9907b69
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to