Jim,

 

I would like to expand my previous reply as follows.

 

JIM> The problem, as I see it, is that the complexity of the level of
general knowledge that is (probably) required for a basic AGI program is too
great for any of these methods to work with anything other than superficial
or at best simple examples.  So if the day came when one method worked then
a great many methods would probably work because it would mean that someone
had figured out how to contain combinatorial complexity or had developed
hardware sufficient to attain minimal AI.

 

SERGIO> This opinion completely ignores the fact that the brain is a
physical system, and physical systems are capable to self-organize
themselves. I know that self-organization is not well understood. I also
know about the controversy surrounding the role of self-organization in
evolution and biology. But these three facts do not mean that
self-organization is not a possible answer to either the structural or the
functional complexity of the brain. The possibility that self-organization
could explain both, is still open. 

 

Of course, one can argue that such a possibility is very remote. But I just
said that self-organization itself, or its role in the anatomy or physiology
of the brain, are not well understood. Under such circumstances, nobody can
say that self-organization does not play a significant role in the evolution
of the brain, just as well as nobody can say that self-organization does
play a significant role in the evolution of the brain. So, I'd say, there
are two options, try to explain the complexity, or try to explain the
self-organization, and both deserve a reasonable chance. 

 

Until very recently, explaining the self-organization was not possible.
Recalling that we are talking about a physical system, there is a principle
in Physics that actually explains self-organization. It says that every
dynamical system that has symmetries, also has a conservation law that
applies to a "conserved quantity." The conserved quantity is something that
is: 1. a property of the system, and 2. remains invariant under the
dynamics. In other words, it is what we call an attractor. There are two
ways to calculate the conserved quantity: Noether's 1918 theorem (and its
many extensions), and my recent work with causal sets. Noether's theorem and
extensions are limited to Lagrangian systems and of little interest in AGI.
My theorem is general, and contains Noether's theorem and extensions (so far
I have proved only one particular case). My theorem says that every causal
system has symmetries and establishes a general procedure to obtain the
attractors. 

 

As a result, I now consider self-organization as fully explained. As
everything, my conclusions are subject to scientific scrutiny, and a long,
arduous process will have to follow to actually apply the theory to a miriad
of particular cases, the brain being only one of them, the GUAPs being
another, biology being a third. This will unify the theory of
self-organization, or so I hope. 

 

To conclude this presentation, I would like to explain why I was so happy
yesterday when I learned about Karl Friston's work. I quote again from
Dynamic causal modelling: A critical review of the biophysical and
statistical foundations
<http://www.fil.ion.ucl.ac.uk/spm/doc/papers/Daunizeau_NeuroImage_58_312_201
1.pdf>   by  J. Daunizeau, O. David, K.E. Stephan, NeuroImage 58 (2011)
312-322:

 

" ... the functional role played by any brain component (e.g., cortical
area, sub-area, neuronal population or neuron) is defined largely by its
connections ... In other terms, function emerges from the flow of
information among brain areas ... effective connectivity refers to causal
effects, i.e., the directed influence that system elements exert on each
other."

 

This statement describes results obtained by neuroscientists from
observation, mostly imaging techniques such as fMRI. My conclusions come
from theoretical considerations about the mathematical properties of causal
sets. And the two are nearly identical. This is strong agreement between
theory and experiment. 

 

At this point, I believe that the second option, that self-organization
plays the largest role in the evolution of the brain, is the most likely one
and needs careful consideration. There is no need to abandon the traditional
"reason and code" approach to AGI. But you need to carefully rethink your
complexity argument. 

 

Sergio

 

PS. I don't mean to disregard the rest of your presentations. I just reply
as I can. 

 

 

From: Jim Bromer [mailto:[email protected]] 
Sent: Thursday, August 16, 2012 6:47 AM
To: AGI
Subject: Re: [agi] Uncertainty, causality, entropy, self-organization, and
Schroedinger's cat.

 

I said:

I have thought about taking actions that can be used to minimize the
complications of numerous unknowns, but since this strategy has to be based
on some method of avoiding the worse outcomes, that means that the strategy
cannot be based on a simplistic way to minimize "entropy".

--------------

 

A probability method is a form of narrow AI. It is based on a presumption
that the probability of what is being measured will be both obvious and
simple.  Neither of these situations are actual realities for AGI
developers.  So simple binary examples examples where the state of the
result can be easily observed, is fine for an introductory explanation but
it is not adequate as a solution to the real problem.

 

AGI and AI programs do not attain the level of child-like intelligence
because many fundamental stages of learning (or of situational analysis) is
too complicated for a contemporary computer program. So while we can imagine
these different methods attaining human-like intelligence, it is not
feasible at this time.  Another variant, like free-action agency is unlikely
to break through that barrier.

 

When combined with a simplistic theory, like the markov process, Sergio's
theory that the physical stratum is the place where the miminalization of
the information entropy is taking place is similar to those theories which
posits that all reasoning takes place through the syntax of expressions.
These kinds of theories have proven inadequate to even model the necessary
types of subject matter.  It is clearly the semantics (or meaning) of
expressions which is of interest to the AGI programmer or AGI dreamer.  So
while we can say that the mind must be developing insight with the 'physics'
of the brain, we must still be able to recognize that it must be doing so on
the basis of the meaning of the objects that the information refers to.  The
attempt to declare that all processes of mind must be reduced to the terms
of purely physical processes is inadequate because it is presumptious
(people do not know exactly how the brain works) and blatently false (a
computer program operates on the basis of its electronic components but at
the same time is is programmable so it can be used to do things that we want
it to do - within limits.  The computer is not just operating on the basis
of the physics of it components but also because people have decided to use
it in certain ways.  This duality of causation may be understood as a result
of other kinds of physical reactions but the theory of everything does not
actually refer to everything but only a frail conjecture about a conjectured
prototype. Without understanding everything we are left with dualities and
multiplicities).

 

It seems to me that the problems of multiplicities is inherently complicated
and these complications will doom many over-simplifications of non-essential
reductionism.

 

Jim Bromer

On Wed, Aug 15, 2012 at 9:55 PM, Jim Bromer <[email protected]> wrote:

Well, I am still sceptical of the theory for one basic reason.  The problem,
as I see it, is that the complexity of the level of general knowledge that
is (probably) required for a basic AGI program is too great for any of these
methods to work with anything other than superficial or at best simple
examples.  So if the day came when one method worked then a great many
methods would probably work because it would mean that someone had figured
out how to contain combinatorial complexity or had developed hardware
sufficient to attain minimal AI.

 

Weighted reasoning was at first believed to be a solution to combinatorial
complexity.  Why didn't it work?  There were a few problems.  One was that
not everything can be expressed in the terms of a range of values.
Secondly, a weighting would, by the very method that it is able to simplify
complex relationships, represent different kinds of things and these
different things get mushed up and produce sub par results. Finally, these
systems have no way to integrate different kinds of relations wisely.

 

Of course, since weighted reasoning is usually used with correlations, one
might imagine that correlation points could be developed to represent when
these complicated relations can be fused and when they should be divided and
how they can be structured and integrated.  This would require some trial
and error methods to learn how to apply these techniques to real world
modelling, but few people actually talk about stuff like this.  And there is
no reason to think that this sort of method can produce intelligence without
first transcending the combinatorial complexity problem.

 

I have thought about taking actions that can be used to minimize the
complications of numerous unknowns, but since this strategy has to be based
on some method of avoiding the worse outcomes, that means that the strategy
cannot be based on a simplistic way to minimize "entropy".  Taking an action
when facing multiple unknowns has to be derived from a biased method that
help the entity avoid the worse outcomes.  This in turn implies that biasing
strategies could also be used in the hope of increasing the chances of
better outcomes based on the projection of insights about the kinds of
situation that the intelligent device thought it might be in.

 

Jim Bromer 

 


AGI |  <https://www.listbox.com/member/archive/303/=now> Archives
<https://www.listbox.com/member/archive/rss/303/18883996-f0d58d57> |
<https://www.listbox.com/member/?&;
ad2> Modify Your Subscription

 <http://www.listbox.com> 

 




-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to