For the discussion of the subject the details of the pattern representation
are not important at all. It is sufficient if you agree that a spoken
sentence represent a certain set of patterns which are translated into the
sentence. The receiving agent retranslates the sentence and matches the
content with its model by activating similar patterns.

The activation of patterns is extremely fast and happens in real time. The
brain even predicts patterns if it just hears the first syllable of a word:

http://www.rochester.edu/news/show.php?id=3244

There is no creation of new patterns and there is no intelligent algorithm
which manipulates patterns. It is just translating, sending, receiving and
retranslating.

>From the ambiguities of natural language you obtain some hints about the
structure of the patterns. But you cannot even expect to obtain all detail
of these patterns by understanding the process of language understanding.
There will be probably many details within these patterns which are only
necessary for internal calculations. 
These details will be not visible from the linguistic point of view. Just
think about communicating computers and you will know what I mean.


- Matthias

Mike Tintner [mailto:[EMAIL PROTECTED] wrote:

Matthias,

You seem - correct me - to be going a long way round saying that words are 
different from concepts - they're just sound-and-letter labels for concepts,

which have a very different form. And the processing of words/language is 
distinct from and relatively simple compared to the processing of the 
underlying concepts.

So take

THE CAT SAT ON THE MAT

or

THE MIND HAS ONLY CERTAIN PARTS WHICH ARE SENTIENT

or

THE US IS THE HOME OF THE FINANCIAL CRISIS

the words "c-a-t" or "m-i-n-d" or "U-S"  or "f-i-n-a-n-c-i-a-l c-r-i-s-i-s" 
are distinct from the underlying concepts. The question is: What form do 
those concepts take? And what is happening in our minds (and what has to 
happen in any mind) when we process those concepts?

You talk of "patterns". What patterns, do you think, form the concept of 
"mind" that are engaged in thinking about sentence 2? Do you think that 
concepts like "mind" or "the US" might involve something much more complex 
still? "Models"? Or is that still way too simple? "Spaces"?

Equally, of course, we can say that each *sentence* above is not just a 
"verbal composition" but a "conceptual composition" - and the question then 
is what form does such a composition take? Do sentences form, say, a 
"pattern of patterns",  or something like a "picture"? Or a "blending of 
spaces" ?

Or are concepts like *money*?

YOU CAN BUY A LOT WITH A MILLION DOLLARS

Does every concept function somewhat like money, e.g. "a million dollars" - 
something that we know can be cashed in, in an infinite variety of ways, but

that we may not have to start "cashing in,"  (when processing), unless 
really called for - or only cash in so far?

P.S. BTW this is the sort of psycho-philosophical discussion that I would 
see as central to AGI, but that most of you don't want to talk about?





Matthias: What the computer makes with the data it receives depends on the 
information
> of the transferred data, its internal algorithms and its internal data.
> This is the same with humans and natural language.
>
>
> Language understanding would be useful to teach the AGI with existing
> knowledge already represented in natural language. But natural language
> understanding suffers from the problem of ambiguities. These ambiguities 
> can
> be solved by having similar knowledge as humans have. But then you have a
> recursive problem because first there has to be solved the problem to 
> obtain
> this knowledge.
>
> Nature solves this problem with embodiment. Different people make similar
> experiences since the laws of nature do not depend on space and time.
> Therefore we all can imagine a dog which is angry. Since we have 
> experienced
> angry dogs but we haven't experienced angry trees we can resolve the
> linguistic ambiguity of my former example and answer the question: Who was
> angry?
>
> The way to obtain knowledge with embodiment is hard and long even in 
> virtual
> worlds.
> If the AGI shall understand natural language it would be necessary that it
> makes similar experiences as humans make in the real world. But this would
> need a very very sophisticated and rich virtual world. At least, there 
> have
> to be angry dogs in the virtual world ;-)
>
> As I have already said I do not think the relation between utility of this
> approach and the costs would be positive for first AGI.
>




-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
https://www.listbox.com/member/?&;
Powered by Listbox: http://www.listbox.com



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34
Powered by Listbox: http://www.listbox.com

Reply via email to