Thanks, Colin.

 

You can add me to your mailing list  J

 

Let me know when you have anything specific to test or evaluate.

 

Best,

 

Peter

 

From: Colin Hales [mailto:[email protected]] 
Sent: Tuesday, May 19, 2015 5:55 PM
To: AGI
Subject: Re: [agi] Institute of General Intelligence (IGI)

 

Hi Peter,

Yep I kind of figured that'd be your take. I know your view of things. This IGI 
is specifically designed to embark on incorporating brain-mimetic natural 
'computation' on substrates. In an IGI version of AGI the role of computers in 
the operation of the intelligence will be determined by the needs of a core 
substrate that forms the physical seat of the AGI. Computation is central. The 
computer as an origin of that computation is an experimental variable. 

 

A summary ....we are looking at the idea that there are 2 fundamental kinds of 
putative AGI (1) & (3), and their hybrid (2) that forms a third approach as 
follows:

 

(1) C-AGI      computer substrate only. Neuromorphic equivalents of it.

(2) H-AGI      hybrid of (1) and (3). The inorganic version is a new kind of 
neuromorphic chip. The organic version has ... erm... organics in it.

(3) S-AGI      synthetic AGI. organic or inorganic. Natural brain physics only. 
No computer. 

 

(aside: S-AGI just came out of my fingers. I hope this is OK, Dorian!)

 

Think this way: What we have now is 100% computer. S-AGI is 100% natural 
physics (organic or inorganic). H-AGI is set somewhere in between.  It's the 
level of computer computation/natural computation that is at issue. All are 
computation.

 

The human brain is a natural version of (3) with a neuronal/astrocyte  
substrate. (3) has no computer whatever in it. it retains all the natural 
physics (whatever that is). H-AGI targets the inclusion of the essential 
natural brain physics in the substrate of (2) and to incorporate (1) 
computer-substrates and software to an extent to be determined. In my case an 
H-AGI would be inorganic. Others see differently. 

 

Where you might have a stake in this?

 

The history of AGI can be summed up as an experiment that seeks to see if the 
role of (1) C-AGI as a brain is fundamentally indistinguishable from (3) S-AGI 
under all conditions. That is the hypothesis. The 65 year old bet that has 
attracted 100% of the investment to date. H-AGI does not make that 
presupposition and seeks to contrast (1) and (3) in revealing ways that then 
allow us to speak authoritatively about the (1)/(3) relationship in AGI 
potential. Only then will we really understand the difference between (1) and 
(3). So far that difference is entirely and intuition. A good one. But only 
intuition. Its time for that intuition to be turned into science. Experiments 
in (1) have ruled to date. Now we seek to do some (2)... E.E. we have 65 years 
of 'control' subject. H-AGI builds the first 'test' subject.

 

How about this? 

 

What would be super cool is if this mighty AGI beast you intend making could be 
turned into the brain of a robot. Then we could contrast what it does with what 
an IGI candidate brain does in an identical robot in the same test. That kind 
of testing vision (as far off as it may seem) is a potential way your work and 
the IGI might interface. Which candidate robot best encounters radical novelty, 
without any human intervention/involvement whatever? .... is a really good 
question. To do this test you'd not need to reveal anything about its workings. 
Observed robot behaviour is decisive.

 

It seems to me that whatever venture you plan, it might be wise to keep an eye 
on any (2)/(3) approaches. IGI or not. Because it is directly informing 
expectations of outcomes in (1). We are currently asking the question "If H-AGI 
were to be championed into existence, what would the first vehicle for that 
look like?" If the enthusiasm maintains it will be sketched into a web page and 
we'll see what it tells us and what to do next. It may halt. It may go. I don't 
know. Worth a shot? You bet.

 

With your (1) C-AGI glasses firmly strapped to your head, your wisdom at all 
stages in this would be well received, whatever the messages. So if you have 
time to keep an  eye on happenings, I for one would appreciate it.

 

regards

 

Colin Hales

 

 

 

On Wed, May 20, 2015 at 6:58 AM, Peter Voss <[email protected]> wrote:

Thanks for asking. Haven’t followed the IGI discussions.  

 

Is this about non-computer based approaches to AGI?  If so, I don’t think I 
have anything positive to contribute.

 

More generally, non-profit orgs need strong focus and champions.  And specific 
goals.

 

From: Benjamin Kapp [mailto:[email protected]] 
Sent: Tuesday, May 19, 2015 12:23 PM
To: AGI
Subject: Re: [agi] Institute of General Intelligence (IGI)

 

Mr. Voss,

Given your understanding of the AGI community do you believe an IGI would be 
redundant?  Would your organization be open to collaborating with the IGI?  Do 
you have any advice for how we could be successful in starting up this 
organization?  Perhaps you would be open to being a member of the board?

 

On Tue, May 19, 2015 at 2:03 PM, Peter Voss <[email protected]> wrote:

Not something that can be adequately covered in a few words, but…. “We’re 
building a fully integrated, top-down & bottom-up, real-time, adaptive 
knowledge (& skill) representation, learning and reasoning engine. We’re using 
a combination of graph representation and NN techniques overlaid with fuzzy, 
adaptive rule systems” – ha!

 

Here again are links for some clues:

 

http://www.kurzweilai.net/essentials-of-general-intelligence-the-direct-path-to-agi
 

http://www.realagi.com/index.html 

https://www.facebook.com/groups/RealAGI/ 

 

 

From: Benjamin Kapp [mailto:[email protected]] 

 

Mr. Voss,

Since you are the founder I'm certain you know what agi-3's methodology is.  In 
a few words (maybe more?) could you share with us what that is?

 

On Tue, May 19, 2015 at 1:24 PM, Peter Voss <[email protected]> wrote:

>http://www.agi-3.com <http://www.agi-3.com/>   They just glue together 
>anything and everything that works.

Actually, no.  We have a very specific theory of AGI and architecture 

Peter Voss

Founder, AGI Innovations Inc. 


AGI |  <https://www.listbox.com/member/archive/303/=now> Archives  
<https://www.listbox.com/member/archive/rss/303/26973278-698fd9ee> |  
<https://www.listbox.com/member/?&;> Modify Your Subscription

 <http://www.listbox.com> 

 


AGI |  <https://www.listbox.com/member/archive/303/=now> Archives  
<https://www.listbox.com/member/archive/rss/303/231420-b637a2b0> |  
<https://www.listbox.com/member/?&;> Modify Your Subscription

 <http://www.listbox.com> 

 


AGI |  <https://www.listbox.com/member/archive/303/=now> Archives  
<https://www.listbox.com/member/archive/rss/303/11721311-20a65d4a> |  
<https://www.listbox.com/member/?&;> Modify Your Subscription

 <http://www.listbox.com> 

 


AGI |  <https://www.listbox.com/member/archive/303/=now> Archives  
<https://www.listbox.com/member/archive/rss/303/231420-b637a2b0> |  
<https://www.listbox.com/member/?&;> Modify Your Subscription

 <http://www.listbox.com> 

 




-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to