Steve,

Tell you what, *we are dangerously off topic*. Let me help you out. I'm not
getting into a pissing match with you on a public listserv. You want to
actually learn something about collecting physiological data and then
drawing conclusions, hey e-mail me at [email protected] and we can spend
the next two weeks going back and forth over what I see as the flaws in
your approach. You really want to go into this with an ultramarathon
runner, sure hey, no problem, but assuming a skeptic and a scientist will
never believe you based on your somewhat limited data? *Hey, bad form
guvnor.* I call shenanigans. Just sayin.

-GJS

On Wed, May 6, 2015 at 2:07 AM, Nanograte Knowledge Technologies <
[email protected]> wrote:

> Hi Colin
>
> You seem to be following a similar process to AI as to what was used to
> develop the first, nuclear bomb - various approaches were used coupled with
> great experimentation.
>
> Semantically, your inclusion of the term "emergent" in your last message
> undersores this approach for me. I'd like to dwell on its relevance for a
> few seconds. Emergence is regarded as the basis for complex-systems
> engineering (Checkland). Further, Checkland asserted how the debate between
> complex and simple systems would probably give rise to what is regarded as
> systems thinking. This is ancient stuff I'm repeating only to stress the
> importance of its credibility. Thus, on the theoretically basis alone, your
> experimental approach could be deemed to be sound.
>
> Narrow AI, broad AI, AGI? All peas in the same pod of complex-systems
> thinking. The fundamentals still have no significant incentive to change.
>
> Personally, I would value such an experimental approach on the basis of
> rethinking the whole idea of developing AI. How else was the sound-barrier
> broken? In addition, if one followed the emerging trend in recent,
> adaptively-autonomous technologies, one would be hard pressed to write off
> your approach.
>
> Just one theoretically-moot point if I may, albeit a semantic one? Any
> institutionalised process effectively is a program of code. As an
> extension, any reduced process - as a procedural implementation - on a
> computer would become a computerized program. Hence, I suppose, your search
> for a generic algorithmic platform.
>
> In the sense of systemically, as soon as you'd link the "stochastic"
> environment to a computer chip in any way it should emerge as a form of
> computer program. Whilst one understands the need for research to be highly
> focussed on its objectives, one must still have a design framework that
> would not unduly restrict any design in a short-sighted
> Heisenbergian-Einstein debate.
>
> I would assume then that you do have a quantum-based design framework
> you're working from. If not though, this particular,  organic approach
> would sooner or later come up against the eco-systemic realities of
> highly-abstracted implementation. This then, mainly due to the lack of
> navigational competency in the R&D framework to consistently and reliably
> perform adaptive integration. If it cannot be measured somehow, it cannot
> be reliably tested and I'm by no means suggesting this to be the case with
> your experiment. Mine are just thoughts on the interesting topic at hand.
> One day, when bootstrapping does occur, you'd be wanting to debug though.
> If only purely mathematical, then purely computational? Maybe that was how
> computer science emerged.
>
> Good luck with the experiment.
>
> Rob
>
> ------------------------------
> To: [email protected]
> From: [email protected]
> Subject: RE: [agi] Re: Starting to Define Algorithms that are More
> Powerfulthan Narrow AI
> Date: Wed, 6 May 2015 10:01:33 +1000
>
> Hi,
> Rather busy... Having trouble devoting time here.
>
> Jim.... You ask if I am making some kind of electric circuit. Basically
> yes. Except it's physical instantiation is important. Materials in space. I
> know you won't get why that is important. That's ok for now. Just accept
> that it's like that for the same reason the brain is like that.
>
> What it isn't is an 'Equivalent circuit' in the traditional sense of
> voltage/current replication. It is designed to produce functionally
> equivalent action potential-style signaling AND the brain-style field
> system that actually expresses the voltages. The hardware will (in the
> field version) express an EEG and MEG   like brains.
>
> Having said that I am currently designing a version that doesn't express
> the fields but allows their addition later...knowing what performance
> degradation results (it will be narrow-AI not AGI). Call it a causality
> mirror with a faked image in it.
>
> It is deeply self modifying. The circuits literally rewire themselves.
> Circuit loops duplicate/diverge and switch out/off. It  accounts for the
> process of brain development as a kind of learning. I.e. I don't even have
> to design the 'brain'. It will self configure based on being in the world.
> Because it's not using neurons it won't automatically mimic brains in
> structure. I have no idea what a brain will look like. Physically its a
> crystalline rock. No actual material growth. Functionally it will stabilize
> in ways I can't know except by experiment. It means that it must be
> permanently juvenile.. Overexpressed neurons and overexpressed synapses
> culled back. Lots of wastage. But so what?
>
> Not one line of software anywhere. Any 'algorithm' it has is in the
> adaptation mechanisms. But they are in hardware. The state of the chip's
> self configuration is the only actual data involved. Yet, when you look at
> it there will be deep regularities in its behaviour. You could write them
> down. However they are all emergent.
>
> You know what the hardest part of this is? ... Giving it goals. A reason
> to bother. A reason for it to sustain the quasi-stable resonances that
> signify its functioning. I have to think of something akin to homeostasis
> to keep it going! ROBEOSTASIS. You know what might happen? It possibly
> self-sustain without human intervention or some kind of hardwiring until
> the fields are added. Unsure. Answering that is an experimental goal. Steve
> seems to be deeply inside homeostatic concerns. So that's good.
>
> I'm not here to justify anything. Experimental proof will speak for me.
> And if I can't get the version with and without the fields to be different
> in predicted ways then I will grovel at the feet of the great god
> computationalism. Not before.[image: Smiling face with smiling eyes]
>
>
> I think the approach is a reversion to 'natural cybernetics' that had a
> brief life in the 1950s and then was lost in a tsunami called computer
> science. I bring it back for an upgrade. Notice that AGI failure started
> the moment cybernetics stopped. The actual science of artificial
> intelligence stopped then, too...IMO.
>
> Enough poking the bear. Gotta get back to it.
>
> I really appreciate the interest in this 'adaptive control' approach.
>
> Cheers
>
> Colin
>
>
> ------------------------------
> From: Jim Bromer <[email protected]>
> Sent: ‎4/‎05/‎2015 12:42 AM
> To: AGI <[email protected]>
> Subject: Re: [agi] Re: Starting to Define Algorithms that are More
> Powerfulthan Narrow AI
>
> I thought the ideas are interesting and Colin's description was more
> readable than usual but the arguments supporting the method weren't
> very powerful.  I am curious about how Colin is implementing the
> method. Could you give me a little more about that? Are you designing
> some kind of electrical circuit?
>
> What I was trying to say in this thread is that you have to supply a
> little more insight about why you think that the methods that you are
> designing and will be implementing would rise above being 'narrow ai'.
> For instance, Colin's honest report on how far he has actually gotten
> so far sounds like it is on par with simple narrow AI. As I reread
> your messages I keep finding a little more in it. But back to my
> point. Since I can rough out the algorithms that I would use as if
> they were abstractions, or as if they could exist within an abstract
> world, it would seem that I should be able to conduct simple tests to
> show that they could diversify in some way that is: 1. at least better
> than narrow ai, and 2. useful in some way. So perhaps I should add
> that. I would say, for example, that artificial neural networks would
> pass this kind of test. However, the criticism then is, ironically
> given our use of the narrow ai term, that they lack efficient means to
> focus and they cannot be efficiently used as componential objects.
>
> So, can you guys define some abstract or simple tests that could show
> that your ideas would become able to adapt to the more complicated
> demands of actual tests? The value of the simple test is that once you
> can get your algorithms to pass the first test you might come up with
> ways to design a slightly more aggressive test. So if I could test my
> ideas to,say, try to learn to recognize some simple classifications
> then I might try to see if I can get it to try to get it to learn to
> utilize systems of classifications effectively and efficiently
> (without redesigning the program only for that specific kind of test.)
> So then I would have to design some other kind of test to make sure
> that it is somewhat general.
> Jim Bromer
>
> On Sun, May 3, 2015 at 3:25 AM, Colin Hales <[email protected]> wrote:
> >
> >
> >> On Sat, May 2, 2015 at 2:50 AM, Steve Richfield <
> [email protected]> wrote:
> >>>
> >>> Jim,
> >>>
> >>> Again, I think I see the POV to solve this. All animals, from single
> cells to us, are fundamentally adaptive process control systems. We use our
> intelligence to live better and more reliably, procreate, etc., much as
> single-celled animals, only with MUCH richer functionality. Everything fits
> this hierarchy of function leading to intelligence.
> >>>
> >>> Then, people like those on this forum start by ignoring this and
> trying to create intelligence from whole cloth. This may be possible, but
> there is NO existence proof for this, no data to guide the effort, etc. In
> short, there is NO reason to expect a whole-cloth approach to work anytime
> during the next century (or two).
> >>>
> >>> However, some of the mathematics of adaptive process control is known,
> and I suspect the rest wouldn't be all that tough - if only SOMEONE were
> working on it.
> >
> >
> > Erm.... guys. This would be me.
> >
> > I am working on it. For well over a decade now. Cognition and
> intelligence is implemented as an adaptive control system replicating,
> inorganically, the natural original called the human (mammal) nervous
> system. I simply replicate it inorganically. Tough job but I am getting
> there. There's no programming. No software. Just radically adaptively
> nested looping processes. In control strategy terms it is a non-stationary
> system (architecture itself is adaptive). Control loops come into existence
> and bifurcate and vanish adaptively. The architecture commences at the
> level of single ion channels and nest at multiple levels that then appear
> in tissue as neurons doing what they do, but need not appear like this in
> the inorganic version. You don't actually need cells at all. These then
> nest at increasing spatiotemporal scales forming coalitions, layers,
> columns and finally whole tissue. All inorganically. All the same at all
> scales from an adaptive control perspective. Power-law scalable. Physically
> and logically.
> >
> > In my case, for the conscious version the hardware includes the
> field-superposing, active additional feedback in the wave mechanics of the
> EM field system produced by brain cells at specific points. The fields form
> an addition/secondary loop modulation that operates orthogonally,
> outside/through the space occupied by the chip substrate.
> >
> > What I am starting with is the 'zombie' or symbolically ungrounded
> version. It doesn't produce the active field system (missing a whole
> control system feedback mechanism) and uses supervised learning
> (externalised by a conscious human trainer) to compensate for the loss of
> the natural role consciousness has as an endogenous supervisor. It will, in
> the zombie form, underperform in precisely the way all computer AGI
> underperforms. This is what is missing when you use computers to do it all.
> You end up with a recipe (software) for pulling Pinocchio's strings.
> Whereas my system bypasses the puppetry altogether. It makes the little
> boy, not the puppet.
> >
> > However you view it, there's nothing else there in a brain except nested
> loops that have power-law responses in two orthogonal axes: sensory and
> cognitive.  Adding the field system to the sensory axis (e.g. visual
> experience) or part of the cognitive axis (e.g. emotional experience)
> provide the active role for consciousness  implemented through the causal
> impact of the Lorentz force within the hardware. I suppose it'd be an
> 'adaptive control loop' philosophy for cognition and 'EM field theory of
> consciousness' combined. No computing needed whatever. Just like the brain.
> Most of the last ten years has been spent figuring out the EM field bits!
> That I am now omitting, knowing what I lose when I do that (i.e.
> consciousness).
> >
> > Teeny weeny Zombie version 0.0 this year I hope. No EM field generation.
> I call it the 'circular causality controller'. I aim to add the EM fields
> later. That part requires $millions. It's chip-foundry stuff.
> >
> > So chalk me in under this 'adaptive control loop' category for AGI
> implementation please. I know this forum is a 'using computers to do AGI'
> forum so I'll just continue to zip it. I haven't mentioned it much over the
> years because it seems that most of you aren't interested in my approach.
> For reference and for the record.... I am the 'AGI as adaptive control' guy.
> >
> > cheers
> > colin
> >
> >>>
> >>>
> >>> I suspect that when the answers are known, it will be a bit like
> spread spectrum communications, where there is a payoff for complexity, but
> where ultimately there is a substitute for designed-in complexity, e.g.
> like the pseudo-random operation of spread spectrum systems. Genetics seems
> to prefer designed-in complexity (like our brains) but there is NO need for
> computers to have such limitations.
> >>>
> >>> Whatever path you take, you must "see a path" to have ANY chance of
> succeeding. You must have a POV that helps you to "cut the crap" in pursuit
> of your goal. Others here are working on whole-cloth approaches, yet
> bristle when challenged for lacking a guiding POV. I see some hope in
> adaptive control math. Perhaps you see something else, but it MUST have an
> associated guiding POV for you to have any hope of succeeding - more than a
> simple list of what it does NOT have.
> >>>
> >>> Steve
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/11721311-f886df0a
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>   *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> |
> Modify <https://www.listbox.com/member/?&;> Your Subscription
> <http://www.listbox.com>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/27055757-c218d4f9> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to