[agi] Re: [agi] P≠NP

2010-08-16 Thread Matt Mahoney
Does anyone have any comments on this proof? I don't have the mathematical 
background to tell if it is correct. But it seems related to the idea from 
algorithmic information theory that the worst case complexity for any algorithm 
is equal to the average case for compressed inputs. Then to show that P != NP 
you would show that SAT (specifically 9-SAT) with compressed inputs has 
exponential average case complexity. That is not quite the approach the paper 
takes, probably because compression is not computable.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Kaj Sotala 
To: agi 
Sent: Thu, August 12, 2010 2:18:13 AM
Subject: [agi] Re: [agi] P≠NP

2010/8/12 John G. Rose 
>
> BTW here is the latest one:
>
> http://www.win.tue.nl/~gwoegi/P-versus-NP/Deolalikar.pdf

See also:

http://www.ugcs.caltech.edu/~stansife/pnp.html - brief summary of the proof

Discussion about whether it's correct:

http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np/
http://rjlipton.wordpress.com/2010/08/09/issues-in-the-proof-that-p?np/
http://rjlipton.wordpress.com/2010/08/10/update-on-deolalikars-proof-that-p≠np/
http://rjlipton.wordpress.com/2010/08/11/deolalikar-responds-to-issues-about-his-p≠np-proof/

http://news.ycombinator.com/item?id=1585850

Wiki page summarizing a lot of the discussion, as well as collecting
many of the links above:

http://michaelnielsen.org/polymath1/index.php?title=Deolalikar%27s_P!%3DNP_paper#Does_the_argument_prove_too_much.3F



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?&;
Powered by Listbox: http://www.listbox.com

* Open Link in New Tab
* Download


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Help requested: Making a list of (non-robotic) AGI low hanging fruit apps

2010-08-07 Thread Matt Mahoney
Wouldn't it depend on the other researcher's area of expertise?

 -- Matt Mahoney, matmaho...@yahoo.com





From: Ben Goertzel 
To: agi 
Sent: Sat, August 7, 2010 9:10:23 PM
Subject: [agi] Help requested: Making a list of (non-robotic) AGI low hanging 
fruit apps

Hi,

A fellow AGI researcher sent me this request, so I figured I'd throw it
out to you guys


I'm putting together an AGI pitch for investors and thinking of low
hanging fruit applications to argue for. I'm intentionally not
involving any mechanics (robots, moving parts, etc.). I'm focusing on
voice (i.e. conversational agents) and perhaps vision-based systems.
Hellen Keller AGI, if you will :)

Along those lines, I'd like any ideas you may have that would fall
under this description. I need to substantiate the case for such AGI
technology by making an argument for high-value apps. All ideas are
welcome.


All serious responses will be appreciated!!

Also, I would be grateful if we
could keep this thread closely focused on direct answers to this
question, rather than
digressive discussions on Helen Keller, the nature of AGI, the definition of AGI
versus narrow AI, the achievability or unachievability of AGI, etc.
etc.  If you think
the question is bad or meaningless or unclear or whatever, that's
fine, but please
start a new thread with a different subject line to make your point.

If the discussion is useful, my intention is to mine the answers into a compact
list to convey to him

Thanks!
Ben G


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?&;
Powered by Listbox: http://www.listbox.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-08-06 Thread Matt Mahoney
Jim, see http://www.scholarpedia.org/article/Algorithmic_probability
I think this answers your questions.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Fri, August 6, 2010 2:18:09 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


I meant:
Did Solomonoff's original idea use randomization to determine the bits of the 
programs that are used to produce the prior probabilities?  I think that the 
answer to that is obviously no.  The randomization of the next bit would used 
in 
the test of the prior probabilities as done using a random sampling.  He 
probably found that students who had some familiarity with statistics would 
initially assume that the prior probability was based on some subset of 
possible 
programs as would be expected from a typical sample, so he gave this statistics 
type of definition to emphasize the extent of what he had in mind.
 
I asked this question just to make sure that I understood what Solomonoff 
Induction was, because Abram had made some statement indicating that I really 
didn't know.  Remember, this particular branch of the discussion was originally 
centered around the question of whether Solomonoff Induction would 
be convergent, even given a way around the incomputability of finding only 
those 
programs that halted.  So while the random testing of the prior probabilities 
is 
of interest to me, I wanted to make sure that there is no evidence that 
Solomonoff Induction is convergent. I am not being petty about this, but I also 
needed to make sure that I understood what Solomonoff Induction is.
 
I am interested in hearing your ideas about your variation of 
Solomonoff Induction because your convergent series, in this context, was 
interesting.
Jim Bromer


On Fri, Aug 6, 2010 at 6:50 AM, Jim Bromer  wrote:

Jim: So, did Solomonoff's original idea involve randomizing whether the next 
bit 
would be a 1 or a 0 in the program? 

Abram: Yep. 

I meant, did Solomonoff's original idea involve randomizing whether the next 
bit 
in the program's that are originally used to produce the prior probabilities 
involve the use of randomizing whether the next bit would be a 1 or a 0?  I 
have 
not been able to find any evidence that it was.  I thought that my question was 
clear but on second thought I guess it wasn't. I think that the part about the 
coin flips was only a method to express that he was interested in the 
probability that a particular string would be produced from all possible 
programs, so that when actually testing the prior probability of a particular 
string the program that was to be run would have to be randomly generated.
Jim Bromer
 
 

 
On Wed, Aug 4, 2010 at 10:27 PM, Abram Demski  wrote:

Jim,
>
>
>Your function may be convergent but it is not a probability. 
>>
>
>True! All the possibilities sum to less than 1. There are ways of addressing 
>this (ie, multiply by a normalizing constant which must also be approximated 
>in 
>a convergent manner), but for the most part adherents of Solomonoff induction 
>don't worry about this too much. What we care about, mostly, is comparing 
>different hyotheses to decide which to favor. The normalizing constant doesn't 
>help us here, so it usually isn't mentioned. 
>
>
>
>
>You said that Solomonoff's original construction involved flipping a coin for 
>the next bit.  What good does that do?
>
>Your intuition is that running totally random programs to get predictions will 
>just produce garbage, and that is fine. The idea of Solomonoff induction, 
>though, is that it will produce systematically less garbage than just flipping 
>coins to get predictions. Most of the garbage programs will be knocked out of 
>the running by the data itself. This is supposed to be the least garbage we 
>can 
>manage without domain-specific knowledge
>
>This is backed up with the proof of dominance, which we haven't talked about 
>yet, but which is really the key argument for the optimality of Solomonoff 
>induction. 
>
>
>
>
>And how does that prove that his original idea was convergent?
>
>The proofs of equivalence between all the different formulations of Solomonoff 
>induction are something I haven't cared to look into too deeply. 
>
>
>
>
>Since his idea is incomputable, there are no algorithms that can be run to 
>demonstrate what he was talking about so the basic idea is papered with all 
>sorts of unverifiable approximations.
>
>I gave you a proof of convergence for one such approximation, and if you wish 
>I 
>can modify it to include a normalizing constant to ensure that it is a 
>probability measure. It would be helpful to me if your criticisms were more 
>specific to that proof.
>
>
>
>So, did Solomonoff's original 

Re: [agi] Epiphany - Statements of Stupidity

2010-08-06 Thread Matt Mahoney
Mike Tintner wrote:
> What will be the SIMPLEST thing that will mark the first sign of AGI ? - 
> Given 
>that there are zero but zero examples of AGI.
 
Machines have already surpassed human intelligence. If you don't think so, try 
this IQ test. http://mattmahoney.net/iq/

Or do you prefer to define intelligence as "more like a human"? In that case I 
agree that AGI will never happen. No machine will ever be more like a human 
than 
a human.

I really don't care how you define it. Either way, computers are profoundly 
affecting the way people interact with each other and with the world. Where is 
the threshold when machines do most of our thinking for us? Who cares as long 
as 
the machines still give us the feeling that we are in charge.

-- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Fri, August 6, 2010 5:57:33 AM
Subject: Re: [agi] Epiphany - Statements of Stupidity


sTEVE:I have  posted plenty about "statements of ignorance", our probable 
inability to  comprehend what an advanced intelligence might be "thinking", 

 
What will be the SIMPLEST thing that will mark the  first sign of AGI ? - Given 
that there are zero but zero examples of  AGI.
 
Don't you think it would be a good idea to begin at  the beginning? With 
"initial AGI"? Rather than "advanced AGI"? 

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] AGI & Int'l Relations

2010-08-03 Thread Matt Mahoney
Steve Richfield wrote:
> Perhaps you failed to note the great disparity between the US and the World's 
>performance since 2003, or that with each year, greater percentages of the GDP 
>is going into fewer and fewer pockets. Kids starting out now don't really have 
>a 
>chance.

Actually the opposite is true. Third world country economies are growing much 
faster than the US and Europe.
http://www.google.com/publicdata?ds=wb-wdi&ctype=l&strail=false&nselm=h&met_y=ny_gdp_mktp_cd&scale_y=lin&ind_y=false&rdim=country&idim=country:AFG:HTI:TZA:UGA:SDN:VNM:NGA&tstart=-31561920&tunit=Y&tlen=48&hl=en&dl=en


>> When you pay people not to work, they are less inclined to work.

> That does NOT explain that there are MANY unemployed for every available job, 
>and that many are falling off the end of their benefits with nothing to help 
>them. This view may have been true long ago, but it is now dated and wrong.

Where is your data? There are also just as many jobs for each unemployed person 
because most people apply for more than one job. The reason for high 
unemployment in the U.S. is that benefits were extended for 2 years and because 
minimum wage laws have made lower paying jobs disappear or move overseas. When 
benefits run out, people will stop holding out for a better job and go back to 
work out of necessity.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Mon, August 2, 2010 10:16:23 PM
Subject: Re: [agi] AGI & Int'l Relations

Matt,


On Mon, Aug 2, 2010 at 1:05 PM, Matt Mahoney  wrote:

Steve Richfield wrote:
>> I would feel a **LOT** better if someone explained SOME scenario to 
>> eventually 
>>emerge from our current economic mess.
>
>
>What economic mess?
>http://www.google.com/publicdata?ds=wb-wdi&ctype=l&strail=false&nselm=h&met_y=ny_gdp_mktp_cd&scale_y=lin&ind_y=false&rdim=country&idim=country:USA&tdim=true&tstart=-31561920&tunit=Y&tlen=48&hl=en&dl=en
>
>
>
Perhaps you failed to note the great disparity between the US and the World's 
performance since 2003, or that with each year, greater percentages of the GDP 
is going into fewer and fewer pockets. Kids starting out now don't really have 
a 
chance.
 

> Unemployment appears to be permanent and getting worse, 
>
>
>When you pay people not to work, they are less inclined to work.
>

That does NOT explain that there are MANY unemployed for every available job, 
and that many are falling off the end of their benefits with nothing to help 
them. This view may have been true long ago, but it is now dated and wrong.

Steve 

 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Walker Lake

2010-08-03 Thread Matt Mahoney
Steve Richfield wrote:
> Disaster scenarios aside, what would YOU have YOUR AGI do to navigate this 
>future?

It won't be my AGI. If it were, I would be a despot and billions of people 
would 
suffer, just like if any other person ruled the world with absolute power. We 
will be much better off if everyone has a voice, and we have an AGI that makes 
that voice available to everyone else.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Mon, August 2, 2010 10:03:27 PM
Subject: Re: [agi] Walker Lake

Matt,


On Mon, Aug 2, 2010 at 1:10 PM, Matt Mahoney  wrote:

Steve Richfield wrote:
>> How about an international ban on the deployment of all unmanned and 
>> automated 
>>weapons?
> 
>How about a ban on suicide bombers to level the playing field?

Of course we already have that. Unfortunately, one begets the other. Hence, we 
seem to have a choice, neither or both. I vote for neither. 



>
>> 1984 has truly arrived.
>
>
>No it hasn't. People want public surveillance.

I'm not sure what you mean by "public" surveillance. Monitoring private phone 
calls? Monitoring otherwise unused web cams? Monitoring your output when you 
use 
the toilet? Where, if anywhere, do YOU draw the line?
 

It is also necessary for AGI. In order for machines to do what you want, they 
have to know what you know.

Unfortunately, knowing everything, any use of this information will either be 
to 
my benefit, or my detriment. Do you foresee any way to limit use to only 
beneficial use?

BTW, decades ago I developed the plan of, when my kids got in some sort of 
trouble in school or elsewhere, to represent their interests as well as 
possible, regardless of whether I agreed with them or not. This worked 
EXTREMELY 
well for me, and for several other families who have tried this. The point is 
that to successfully represent their interests, I had to know what was 
happening. Potential embarrassment and explainability limited the kids' 
actions. 
I wonder if the same would work for AGIs?
 

In order for a global brain to use that knowledge, it has to be public.

Again, where do you draw the line between public and private?
 

AGI has to be a global  brain because it is too expensive to build any other 
way, and because it would be too dangerous if the whole world didn't control it.

I'm not sure what you mean by "control". 


Here is the BIG question in my own mind, that I have asked in various ways, so 
far without any recognizable answer:

There are plainly lots of things wrong with our society. We got here by doing 
what we wanted, and by having our representatives do what we wanted them to do. 
Clearly some social re-engineering is in our future, if we are to thrive in the 
foreseeable future. All changes are resisted by some, and I suspect that some 
needed changes will be resisted by most, and perhaps nearly everyone. Disaster 
scenarios aside, what would YOU have YOUR AGI do to navigate this future?

To help guide your answer, I see that the various proposed systems of "ethics" 
would prevent breaking the eggs needed to make a good futuristic omelet. I 
suspect that completely democratic systems have run their course. Against this 
is "letting AGI loose" has its own unfathomable hazards. I've been hanging 
around here for quite a while, and I don't yet see any "success path" to work 
toward.

I'm on your side in that any successful AGI would have to have the information 
and the POWER to succeed, akin to Colossus, the Forbin Project, which I 
personally see as more of a success story than a horror scenario. Absent that, 
AGIs will only add to our present problems.

What is the "success path" that you see?

Steve


 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Walker Lake

2010-08-03 Thread Matt Mahoney
Samantha Atkins wrote:
>> No it hasn't. People want public surveillance.
> Guess I am not people then. 

Then why are you posting your response to a public forum instead of replying by 
encrypted private email? People want their words to be available to the world.

> I don't think the global brain needs to know exactly how often I have sex or 
>with whom or in what varieties.  Do you?  

A home surveillance system needs to know who is in your house and whether they 
belong there. If it is intelligent then it will know that you prefer not to 
have 
video of you having sex broadcast on the internet. At the same time, it has to 
recognize what you are doing.

Public surveillance is less objectionable because it will be two-way and can't 
be abused. If someone searches for information about you, then you get a 
notification of who it was and what they learned. I describe how this works 
in http://mattmahoney.net/agi2.html

> No humans will control it and it is not going to be that expensive.

Humans will eventually lose control of anything that is smarter than them. But 
we should delay that as long as possible by making the required threshold the 
organized intelligence of all humanity, and make that organization as efficient 
as possible. The cost is on the order of $1 quadrillion because the knowledge 
that AGI needs is mostly in billions of human brains and there is no quick way 
to extract it.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Samantha Atkins 
To: agi 
Sent: Tue, August 3, 2010 6:49:34 AM
Subject: Re: [agi] Walker Lake

Matt Mahoney wrote: 
Steve Richfield wrote:
>> How about an international ban on the deployment of all unmanned and 
>> automated 
>>weapons?
> 
>How about a ban on suicide bombers to level the playing field?
>
>
>> 1984 has truly arrived.
>
>
>No it hasn't. People want public surveillance. 
Guess I am not people then.  Actually I think surveillance is inevitable given 
current and all but certain future tech.  However, I recognize that human 
beings 
today, and especially their governments, are not remotely ready for it.   To be 
ready for it at the very least the State would have to consider a great number 
of things none of its business to attempt to legislate for or against.  As it 
is 
with the current incredible number of arcane laws on the books it would be very 
easy to see the already ridiculously large prison population of the US double.  
 
Also, please note that full surveillance means no successful rebellion no 
matter 
how bad the powers that be become and how ineffectual the means that let remain 
legal are to change things.  Ever. 



It is also necessary for AGI. In order for machines to do what you want, they 
have to know what you know. 

It is not necessary to have every waking moment surveilled in order to have AGI 
know what we want. 



In order for a global brain to use that knowledge, it has to be public. 
I don't think the global brain needs to know exactly how often I have sex or 
with whom or in what varieties.  Do you?  



AGI has to be a global brain because it is too expensive to build any other 
way, 
and because it would be too dangerous ifthe whole world didn't control it.
No humans will control it and it is not going to be that expensive.

- samantha


agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Shhh!

2010-08-02 Thread Matt Mahoney
Jim, you are thinking out loud. There is no such thing as "trans-infinite". How 
about posting when you actually solve the problem.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Mon, August 2, 2010 9:06:53 AM
Subject: [agi] Re: Shhh!


I think I can write an abbreviated version, but there would only be a few 
people 
in the world who would both believe me and understand why it would work.


On Mon, Aug 2, 2010 at 8:53 AM, Jim Bromer  wrote:

I can write an algorithm that is capable of describing ('reaching') every 
possible irrational number - given infinite resources.  The infinite is not a 
number-like object, it is an active form of incrementation or concatenation.  
So 
I can write an algorithm that can write every finite state of every possible 
number.  However, it would take another algorithm to 'prove' it.  Given an 
irrational number, this other algorithm could find the infinite incrementation 
for every digit of the given number.  Each possible number (including 
the incrementation of those numbers that cannot be represented in truncated 
form) is embedded within a single infinite infinite incrementation of digits 
that is produced by the algorithm, so the second algorithm would have to 
calculate where you would find each digit of the given irrational number by 
increment.  But the thing is, both functions would be computable and provable.  
(I haven't actually figured the second algorithm out yet, but it is not 
a difficult problem.)
> 
>This means that the Trans-Infinite Is Computable.  But don't tell anyone about 
>this, it's a secret.
> 

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Walker Lake

2010-08-02 Thread Matt Mahoney
Steve Richfield wrote:
> How about an international ban on the deployment of all unmanned and 
> automated 
>weapons?
 
How about a ban on suicide bombers to level the playing field?

> 1984 has truly arrived.

No it hasn't. People want public surveillance. It is also necessary for AGI. In 
order for machines to do what you want, they have to know what you know. In 
order for a global brain to use that knowledge, it has to be public. AGI has to 
be a global brain because it is too expensive to build any other way, and 
because it would be too dangerous if the whole world didn't control it.

-- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Mon, August 2, 2010 10:40:20 AM
Subject: [agi] Walker Lake

Sometime when you are flying between the northwest US to/from Las Vegas, look 
out your window as you fly over Walker Lake in eastern Nevada. At the south end 
you will see a system of roads leading to tiny buildings, all surrounded by 
military security. From what I have been able to figure out, you will find the 
U.S. arsenal of chemical and biological weapons housed there. No, we are not 
now 
making these weapons, but neither are we disposing of them.

Similarly, there has been discussion of developing advanced military technology 
using AGI and other computer-related methods. I believe that these efforts are 
fundamentally anti-democratic, as they allow a small number of people to 
control 
a large number of people. Gone are the days when people voted with their 
swords. 
We now have the best government that money can buy monitoring our every email, 
including this one, to identify anyone resisting such efforts. 1984 has truly 
arrived. This can only lead to a horrible end to freedom, with AGIs doing their 
part and more.

Like chemical and biological weapons, unmanned and automated weapons should be 
BANNED. Unfortunately, doing so would provide a window of opportunity for 
others 
to deploy them. However, if we make these and stick them in yet another 
building 
at the south end of Walker Lake, we would be ready in case other nations deploy 
such weapons.

How about an international ban on the deployment of all unmanned and automated 
weapons? The U.S. won't now even agree to ban land mines. At least this would 
restore SOME relationship between popular support and military might. Doesn't 
it 
sound "ethical" to insist that a human being decide when to end another human 
being's life? Doesn't it sound "fair" to require the decision maker to be in 
harm's way, especially when the person being killed is in or around their own 
home? Doesn't it sound unethical to add to the present situation? When deployed 
on a large scale, aren't these WMDs?
 
Steve

 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] AGI & Int'l Relations

2010-08-02 Thread Matt Mahoney
Steve Richfield wrote:
> I would feel a **LOT** better if someone explained SOME scenario to 
> eventually 
>emerge from our current economic mess.

What economic mess?
http://www.google.com/publicdata?ds=wb-wdi&ctype=l&strail=false&nselm=h&met_y=ny_gdp_mktp_cd&scale_y=lin&ind_y=false&rdim=country&idim=country:USA&tdim=true&tstart=-31561920&tunit=Y&tlen=48&hl=en&dl=en


> Unemployment appears to be permanent and getting worse, 

When you pay people not to work, they are less inclined to work.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Mon, August 2, 2010 11:54:25 AM
Subject: Re: [agi] AGI & Int'l Relations

Jan

I can see that I didn't state one of my points clearly enough...


On Sun, Aug 1, 2010 at 3:04 PM, Jan Klauck  wrote:


>
>> My simple (and completely unacceptable) cure for this is to tax savings,
>> to force the money back into the economy.
>
>You have either consumption or savings. The savings are put back into
>the economy in form of credits to those who invest the money.
>

Our present economic problem is that those "credits" aren't being turned over 
fast enough to keep the economic engine running well. At present, with present 
systems in place, there is little motivation to quickly turn over one's wealth, 
and lots of motivation to very carefully protect it. The result is that most of 
the wealth of the world is just sitting there in various accounts, and is NOT 
being spent/invested on various business propositions to benefit the population 
of the world.

We need to do SOMETHING to get the wealth out of the metaphorical mattresses 
and 
back into the economy. Taxation is about the only effective tool that the 
government hasn't already dulled beyond utility. However, taxation doesn't 
stand 
a chance without the cooperation of other countries to do the same. There seems 
to be enough lobbying power in the hands of those with the money to stop any 
such efforts, or at least to leave enough safe havens to make such efforts 
futile.

I would feel a **LOT** better if someone explained SOME scenario to eventually 
emerge from our current economic mess. Unemployment appears to be permanent and 
getting worse, as does the research situation. All I hear are people citing 
stock prices and claiming that the economy is turning around, when I see little 
connection between stock prices and on-the-street economy.

This is an IR problems of monumental proportions. What would YOU do about it?

Steve



 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] AGI & Alife

2010-07-30 Thread Matt Mahoney
Ian Parker wrote

>> Then define your political objectives. No holes, no ambiguity, no
>> forgotten cases. Or does the AGI ask for our feedback during mission?
>> If yes, down to what detail?
>
> With Matt's ideas it does exactly that.

Well, no it doesn't. My proposed AGI facilitates communication between people 
by 
storing and publishing your communication, authenticating it, indexing it, 
organizing it, rating it, providing redundant backup, making it searchable, and 
routing it to anyone who cares. The G in AGI comes from having access to lots 
of 
narrow AI specialists built using existing technology. The specialists come 
from 
the economic incentive to provide quality information in return for attention 
and reputation.

If you define winning a war as achieving your political objectives, then it is 
clear that Al Qaida has defeated the U.S. But my AGI is not going to fight your 
war for you. I think it will prevent wars because it will make it hard to keep 
secrets and it will give you better alternatives to solving your problems than 
killing people.

 -- Matt Mahoney, matmaho...@yahoo.com



- Original Message 
From: Jan Klauck 
To: agi 
Sent: Fri, July 30, 2010 7:25:06 PM
Subject: Re: [agi] AGI & Alife

Ian Parker wrote

>> Then define your political objectives. No holes, no ambiguity, no
>> forgotten cases. Or does the AGI ask for our feedback during mission?
>> If yes, down to what detail?
>
> With Matt's ideas it does exactly that.

How does it know when to ask? You give it rules, but those rules can
be somehow imperfect. How are its actions monitored and sanctioned?
And hopefully it's clear that we are now far from mathematical proof.

> No we simply add to the axiom pool.

Adding is simple, proving is not. Especially when the rules, goals,
and constraints are not arithmetic but ontological and normative
statements. Wether by NL or formal system, it's error-prone to
specify our knowledge of the world (much of it is implicit) and
teach it to the AGI. It's similar to law which is similar to math
with referenced axioms and definitions and a substitution process.
You often find flaws--most are harmless, some are not.

Proofs give us islands of certainty in an explored sea within the
ocean of the possible. We end up with heuristics. That's what this
discussion is about, when I remember right. :)

cu Jan


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?&;
Powered by Listbox: http://www.listbox.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] AGI & Alife

2010-07-28 Thread Matt Mahoney
Ian Parker wrote:
> Matt Mahoney has costed his view of AGI. I say that costs must be recoverable 
>as we go along. Matt, don't frighten people with a high estimate of cost. 
>Frighten people instead with the bill they are paying now for dumb systems.

It is not my intent to scare people out of building AGI, but rather to be 
realistic about its costs. Building machines that do what we want is a much 
harder problem than building intelligent machines. Machines surpassed human 
intelligence 50 years ago. But getting them to do useful work is still a $60 
trillion per year problem. It's going to happen, but not as quickly as one 
might 
hope.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Ian Parker 
To: agi 
Sent: Wed, July 28, 2010 6:54:05 AM
Subject: Re: [agi] AGI & Alife




On 27 July 2010 21:06, Jan Klauck  wrote:

>
>> Second observation about societal punishment eliminating free loaders. The
>> fact of the matter is that "*freeloading*" is less of a problem in
>> advanced societies than misplaced unselfishness.
>
>Fact of the matter, hm? Freeloading is an inherent problem in many
>social configurations. 9/11 brought down two towers, freeloading can
>bring down an entire country.
>
>
>
There are very considerable knock on costs. There is the mushrooming cost of 
security  This manifests itself in many ways. There is the cost of disruption 
to 
air travel. If someone rides on a plane without a ticket no one's life is put 
at 
risk. There are the military costs, it costs $1m per year to keep a soldier in 
Afghanistan. I don't know how much a Taliban fighter costs, but it must be a 
lot 
less.

Clearly any reduction in these costs would be welcomed. If someone were to come 
along in the guise of social simulation and offer a reduction in these costs 
the 
research would pay for itself many times over. "What you are interested in.

This may be a somewhat unpopular thing to say, but money is important. Matt 
Mahoney has costed his view of AGI. I say that costs must be recoverable as we 
go along. Matt, don't frighten people with a high estimate of cost. Frighten 
people instead with the bill they are paying now for dumb systems.
 
> simulations seem :-
>>
>> 1) To be better done by Calculus.
>
>You usually use both, equations and heuristics. It depends on the
>problem, your resources, your questions, the people working with it
>a.s.o.
>

That is the way things should be done. I agree absolutely. We could in fact 
take 
steepest descent (Calculus) and GAs and combine them together in a single 
composite program. This would in fact be quite a useful exercise. We would also 
eliminate genes that simply dealt with Calculus and steepest descent.

I don't know whether it is useful to think in topological terms.


  - Ian Parker
 

>
>---
>agi
>Archives: https://www.listbox.com/member/archive/303/=now
>RSS Feed: https://www.listbox.com/member/archive/rss/303/
>Modify Your Subscription: https://www.listbox.com/member/?&;
>Powered by Listbox: http://www.listbox.com
>

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-24 Thread Matt Mahoney
Mike Tintner wrote:
> Which is?
 
The one right behind your eyes.

-- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Sat, July 24, 2010 9:00:42 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI


Matt: 
I mean a neural model with increasingly complex features, as opposed to an  
algorithmic 3-D model (like video game graphics in reverse). Of course David  
rejects such ideas ( http://practicalai.org/Prize/Default.aspx )  even though 
the one proven working vision model uses it.
 
 
Which is? and does what?  (I'm starting to consider that vision and  visual 
perception  -  or perhaps one should say "common sense", since  no sense in 
humans works independent of the others -  may well be  considerably *more* 
complex than language. The evolutionary time required to  develop our common 
sense perception and conception of the world was vastly  greater than that 
required to develop language. And we are as a culture merely  in our babbling 
infancy in beginning to understand how sensory images work and  are processed).
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-24 Thread Matt Mahoney
Mike Tintner wrote:
> Huh, Matt? What examples of this "holistic" scene analysis are there (or are 
>you thinking about)?
 
I mean a neural model with increasingly complex features, as opposed to an 
algorithmic 3-D model (like video game graphics in reverse).

Of course David rejects such ideas ( http://practicalai.org/Prize/Default.aspx 
) 
even though the one proven working vision model uses it.

-- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Sat, July 24, 2010 6:16:07 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI


Huh, Matt? What examples of this "holistic" scene  analysis are there (or are 
you thinking about)?


From: Matt Mahoney 
Sent: Saturday, July 24, 2010 10:25 PM
To: agi 
Subject: Re: [agi] Re: Huge Progress on the Core of  AGI

David Jones wrote:
> I should also mention that I ran into  problems mainly because I was having a 
>hard time deciding how to identify  objects and determine what is really going 
>on in a scene.

I think that your approach makes the problem harder than it needs to be  (not 
that it is easy). Natural language processing is hard, so researchers in an  
attempt to break down the task into simpler parts, focused on steps like 
lexical  
analysis, parsing, part of speech resolution, and semantic analysis. While 
these  
problems went unsolved, Google went directly to a solution by skipping  them.

Likewise, parsing an image into physically separate objects and then  building 
a 
3-D model makes the problem harder, not easier. Again, look at the  whole 
picture. You input an image and output a response. Let the system figure  out 
which features are important. If your goal is to count basketball passes,  then 
it is irrelevant whether the AGI recognizes that somebody is wearing a  gorilla 
suit.

 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-24 Thread Matt Mahoney
Jim Bromer wrote:
> Solomonoff Induction may require a trans-infinite level of complexity just to 
>run each program. 

"Trans-infinite" is not a mathematically defined term as far as I can tell. 
Maybe you mean larger than infinity, as in the infinite set of real numbers is 
larger than the infinite set of natural numbers (which is true).

But it is not true that Solomonoff induction requires more than aleph-null 
operations. (Aleph-null is the size of the set of natural numbers, the 
"smallest 
infinity"). An exact calculation requires that you test aleph-null programs for 
aleph-null time steps each. There are aleph-null programs because each program 
is a finite length string, and there is a 1 to 1 correspondence between the set 
of finite strings and N, the set of natural numbers. Also, each program 
requires 
aleph-null computation in the case that it runs forever, because each step in 
the infinite computation can be numbered 1, 2, 3...

However, the total amount of computation is still aleph-null because each step 
of each program can be described by an ordered pair (m,n) in N^2, meaning the 
n'th step of the m'th program, where m and n are natural numbers. The 
cardinality of N^2 is the same as the cardinality of N because there is a 1 to 
1 
correspondence between the sets. You can order the ordered pairs as (1,1), 
(1,2), (2,1), (1,3), (2,2), (3,1), (1,4), (2,3), (3,2), (4,1), (1,5), etc. 
See http://en.wikipedia.org/wiki/Countable_set#More_formal_introduction

Furthermore you may approximate Solomonoff induction to any desired precision 
with finite computation. Simply interleave the execution of all programs as 
indicated in the ordering of ordered pairs that I just gave, where the programs 
are ordered from shortest to longest. Take the shortest program found so far 
that outputs your string, x. It is guaranteed that this algorithm will approach 
and eventually find the shortest program that outputs x given sufficient time, 
because this program exists and it halts.

In case you are wondering how Solomonoff induction is not computable, the 
problem is that after this algorithm finds the true shortest program that 
outputs x, it will keep running forever and you might still be wondering if a 
shorter program is forthcoming. In general you won't know.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Sat, July 24, 2010 3:59:18 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


Solomonoff Induction may require a trans-infinite level of complexity just to 
run each program.  Suppose each program is iterated through the enumeration of 
its instructions.  Then, not only do the infinity of possible programs need to 
be run, many combinations of the infinite programs from each simulated Turing 
Machine also have to be tried.  All the possible combinations of (accepted) 
programs, one from any two or more of the (accepted) programs produced by each 
simulated Turing Machine, have to be tried.  Although these combinations of 
programs from each of the simulated Turing Machine may not all be unique, they 
all have to be tried.  Since each simulated Turing Machine would produce 
infinite programs, I am pretty sure that this means that Solmonoff Induction 
is, 
by definition, trans-infinite.
Jim Bromer


On Thu, Jul 22, 2010 at 2:06 PM, Jim Bromer  wrote:

I have to retract my claim that the programs of Solomonoff Induction would be 
trans-infinite.  Each of the infinite individual programs could be enumerated 
by 
their individual instructions so some combination of unique individual programs 
would not correspond to a unique program but to the enumerated program that 
corresponds to the string of their individual instructions.  So I got that one 
wrong.
>Jim Bromer

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-24 Thread Matt Mahoney
David Jones wrote:
> I should also mention that I ran into problems mainly because I was having a 
>hard time deciding how to identify objects and determine what is really going 
>on 
>in a scene.

I think that your approach makes the problem harder than it needs to be (not 
that it is easy). Natural language processing is hard, so researchers in an 
attempt to break down the task into simpler parts, focused on steps like 
lexical 
analysis, parsing, part of speech resolution, and semantic analysis. While 
these 
problems went unsolved, Google went directly to a solution by skipping them.

Likewise, parsing an image into physically separate objects and then building a 
3-D model makes the problem harder, not easier. Again, look at the whole 
picture. You input an image and output a response. Let the system figure out 
which features are important. If your goal is to count basketball passes, then 
it is irrelevant whether the AGI recognizes that somebody is wearing a gorilla 
suit.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Sat, July 24, 2010 2:25:49 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI

Abram,

I should also mention that I ran into problems mainly because I was having a 
hard time deciding how to identify objects and determine what is really going 
on 
in a scene. This adds a whole other layer of complexity to hypotheses. It's not 
just about what is more predictive of the observations, it is about deciding 
what exactly you are observing in the first place. (although you might say its 
the same problem).

I ran into this problem when my algorithm finds matches between items that are 
not the same. Or it may not find any matches between items that are the same, 
but have changed. So, how do you decide whether it is 1) the same object, 2) a 
different object or 3) the same object but it has changed. 

And how do you decide its relationship to something else...  is it 1) 
dependently attached 2) semi-dependently attached(can move independently, but 
only in certain ways. Yet also moves dependently) 3) independent 4) sometimes 
dependent 5) was dependent, but no longer is, 6) was dependent on something 
else, but then was independent, but now is dependent on something new. 


These hypotheses are different ways of explaining the same observations, but 
are 
complicated by the fact that we aren't sure of the identity of the objects we 
are observing in the first place. Multiple hypotheses may fit the same 
observations, and its hard to decide why one is simpler or better than the 
other. The object you were observing at first may have disappeared. A new 
object 
may have appeared at the same time (this is why screenshots are a bit 
malicious). Or the object you were observing may have changed. In screenshots, 
sometimes the objects that you are trying to identify as different never appear 
at the same time because they always completely occlude each other. So, that 
can 
make it extremely difficult to decide whether they are the same object that has 
changed or different objects.

Such ambiguities are common in AGI. It is unclear to me yet how to deal with 
them effectively, although I am continuing to work hard on it. 


I know its a bit of a mess, but I'm just trying to demonstrate the trouble I've 
run into. 


I hope that makes it more clear why I'm having so much trouble finding a way of 
determining what hypothesis is most predictive and simplest.

Dave


On Thu, Jul 22, 2010 at 10:23 PM, Abram Demski  wrote:

David,
>
>What are the different ways you are thinking of for measuring the 
>predictiveness? I can think of a few different possibilities (such as 
>measuring 
>number incorrect vs measuring fraction incorrect, et cetera) but I'm wondering 
>which variations you consider significant/troublesome/etc.
>
>--Abram
>
>
>On Thu, Jul 22, 2010 at 7:12 PM, David Jones  wrote:
>
>It's certainly not as simple as you claim. First, assigning a probability is 
>not 
>always possible, nor is it easy. The factors in calculating that probability 
>are 
>unknown and are not the same for every instance. Since we do not know what 
>combination of observations we will see, we cannot have a predefined set of 
>probabilities, nor is it any easier to create a probability function that 
>generates them for us. That is just as exactly what I meant by quantitatively 
>define the predictiveness... it would be proportional to the probability. 
>
>>Second, if you can define a program ina way that is always simpler when it is 
>>smaller, then you can do the same thing without a program. I don't think it 
>>makes any sense to do it this way. 
>>
>>It is not that simple. If it was, we could solve a large portion of agi 
easily.
>>On Thu, Jul 22, 2010 at 3:16 PM, Matt Mahoney  wrote:
>

Re: [agi] Pretty worldchanging

2010-07-23 Thread Matt Mahoney
The video says it has 2 GB of memory. I assume that's SSD and there is no disk.

It's actually not hard to find a computer for $35. People are always throwing 
away old computers that still work.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Fri, July 23, 2010 9:50:44 AM
Subject: [agi] Pretty worldchanging


this strikes me as socially worldchanging if it  works - potentially leading to 
you-ain't-see-nothing-yet changes in world  education (& commerce) levels over 
the next decade:
 
http://www.physorg.com/news199083092.html
 
Any comments on its technical & massproduction  viability ?
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-22 Thread Matt Mahoney
Jim Bromer wrote:
> Please give me a little more explanation why you say the fundamental method 
> is 
>that the probability of a string x is proportional to the sum of all programs 
>M 
>that output x weighted by 2^-|M|.  Why is the M in a bracket?

By |M| I mean the length of the program M in bits. Why 2^-|M|? Because each bit 
means you can have twice as many programs, so they should count half as much.

Being uncomputable doesn't make it wrong. The fact that there is no general 
procedure for finding the shortest program that outputs a string doesn't mean 
that you can never find it, or that for many cases you can't approximate it.

You apply Solomonoff induction all the time. What is the next bit in these 
sequences?

1. 0101010101010101010101010101010

2. 11001001110110101010001

In sequence 1 there is an obvious pattern with a short description. You can 
find 
a short program that outputs 0 and 1 alternately forever, so you predict the 
next bit will be 1. It might not be the shortest program, but it is enough that 
"alternate 0 and 1 forever" is shorter than "alternate 0 and 1 15 times 
followed 
by 00" that you can confidently predict the first hypothesis is more likely.

The second sequence is not so obvious. It looks like random bits. With enough 
intelligence (or computation) you might discover that the sequence is a binary 
representation of pi, and therefore the next bit is 0. But the fact that you 
might not discover the shortest description does not invalidate the principle. 
It just says that you can't always apply Solomonoff induction and get the 
number 
you want.

Perhaps http://en.wikipedia.org/wiki/Kolmogorov_complexity will make this clear.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Thu, July 22, 2010 5:06:12 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


On Wed, Jul 21, 2010 at 8:47 PM, Matt Mahoney  wrote:
The fundamental method is that the probability of a string x is proportional to 
the sum of all programs M that output x weighted by 2^-|M|. That probability is 
dominated by the shortest program, but it is equally uncomputable either way.
Also, please point me to this mathematical community that you claim rejects 
Solomonoff induction. Can you find even one paper that refutes it?
 
You give a precise statement of the probability in general terms, but then say 
that it is uncomputable.  Then you ask if there is a paper that refutes it.  
Well, why would any serious mathematician bother to refute it since you 
yourself 
acknowledge that it is uncomputable and therefore unverifiable and therefore 
not 
a mathematical theorem that can be proven true or false?  It isn't like you 
claimed that the mathematical statement is verifiable. It is as if you are 
making a statement and then ducking any responsibility for it by denying that 
it 
is even an evaluation.  You honestly don't see the irregularity?
 
My point is that the general mathematical community doesn't accept Solomonoff 
Induction, not that I have a paper that "refutes it," whatever that would mean.
 
Please give me a little more explanation why you say the fundamental method is 
that the probability of a string x is proportional to the sum of all programs M 
that output x weighted by 2^-|M|.  Why is the M in a bracket?

 
On Wed, Jul 21, 2010 at 8:47 PM, Matt Mahoney  wrote:

Jim Bromer wrote:
>> The fundamental method of Solmonoff Induction is trans-infinite.
>
>
>The fundamental method is that the probability of a string x is proportional 
>to 
>the sum of all programs M that output x weighted by 2^-|M|. That probability 
>is 
>dominated by the shortest program, but it is equally uncomputable either way. 
>How does this approximation invalidate Solomonoff induction?
>
>
>Also, please point me to this mathematical community that you claim rejects 
>Solomonoff induction. Can you find even one paper that refutes it?
>
> -- Matt Mahoney, matmaho...@yahoo.com 
>
>
>
>
>

 From: Jim Bromer 
>To: agi 
>Sent: Wed, July 21, 2010 3:08:13 PM 
>
>Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction
> 
>
>
>I should have said, It would be unwise to claim that this method could stand 
>as 
>an "ideal" for some valid and feasible application of probability.
>Jim Bromer
>
>
>On Wed, Jul 21, 2010 at 2:47 PM, Jim Bromer  wrote:
>
>The fundamental method of Solmonoff Induction is trans-infinite.  Suppose you 
>iterate through all possible programs, combining different programs as you go. 
> 
>Then you have an infinite number of possible programs which have a 
>trans-infinite number of combinations, because each tier of combinations can 
>then be recombined to produce a second, 

Re: [agi] How do we hear music

2010-07-22 Thread Matt Mahoney
deepakjnath wrote:

> Why do we listen to a song sung in different scale and yet identify it as the 
>same song.?  Does it have something to do with the fundamental way in which we 
>store memory?

For the same reason that gray looks green on a red background. You have more 
neurons that respond to differences in tones than to absolute frequencies.

 -- Matt Mahoney, matmaho...@yahoo.com





From: deepakjnath 
To: agi 
Sent: Thu, July 22, 2010 3:59:57 PM
Subject: [agi] How do we hear music

Why do we listen to a song sung in different scale and yet identify it as the 
same song.?  Does it have something to do with the fundamental way in which we 
store memory?

cheers,
Deepak

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-22 Thread Matt Mahoney
David Jones wrote:
> But, I am amazed at how difficult it is to quantitatively define more 
>predictive and simpler for specific problems. 

It isn't hard. To measure predictiveness, you assign a probability to each 
possible outcome. If the actual outcome has probability p, you score a penalty 
of log(1/p) bits. To measure simplicity, use the compressed size of the code 
for 
your prediction algorithm. Then add the two scores together. That's how it is 
done in the Calgary challenge http://www.mailcom.com/challenge/ and in my own 
text compression benchmark.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Thu, July 22, 2010 3:11:46 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI

Because simpler is not better if it is less predictive.



On Thu, Jul 22, 2010 at 1:21 PM, Abram Demski  wrote:

Jim,
>
>Why more predictive *and then* simpler?
>
>--Abram
>
>
>On Thu, Jul 22, 2010 at 11:49 AM, David Jones  wrote:
>
>An Update
>>
>>I think the following gets to the heart of general AI and what it takes to  
>>achieve it. It also provides us with evidence as to why general AI is so 
>>difficult. With this new knowledge in mind, I think I will be much more 
>>capable 
>>now  of solving the problems and making it work. 
>>
>>
>>I've come to the conclusion lately that the best hypothesis is better because 
>>it 
>>is more predictive and then simpler than other hypotheses (in that order 
>>more predictive... then simpler). But, I am amazed at how difficult it is to 
>>quantitatively define more predictive and simpler for specific problems. This 
>>is 
>>why I have sometimes doubted the truth of the statement.
>>
>>In addition, the observations that the AI gets are not representative of all 
>>observations! This means that if your measure of "predictiveness" depends on 
>>the 
>>number of certain observations, it could make mistakes! So, the specific 
>>observations you are aware of may be unrepresentative of the predictiveness 
>>of a 
>>hypothesis relative to the truth. If you try to calculate which hypothesis is 
>>more predictive and you don't have the critical observations that would give 
>>you 
>>the right answer, you may get the wrong answer! This all depends of course on 
>>your method of calculation, which is quite elusive to define. 
>>
>>
>>Visual input from screenshots, for example, can be somewhat malicious. Things 
>>can move, appear, disappear or occlude each other suddenly. So, without 
>>sufficient knowledge it is hard to decide whether matches you find between 
>>such 
>>large changes are because it is the same object or a different object. This 
>>may 
>>indicate that bias and preprogrammed experience should be introduced to the 
>>AI 
>>before training. Either that or the training inputs should be carefully 
>>chosen 
>>to avoid malicious input and to make them nice for learning. 
>>
>>
>>This is the "correspondence problem" that is typical of computer vision and 
>>has 
>>never been properly solved. Such malicious input also makes it difficult to 
>>learn automatically because the AI doesn't have sufficient experience to know 
>>which changes or transformations are acceptable and which are not. It is 
>>immediately bombarded with malicious inputs.
>>
>>I've also realized that if a hypothesis is more "explanatory", it may be 
>>better. 
>>But quantitatively defining explanatory is also elusive and truly depends on 
>>the 
>>specific problems you are applying it to because it is a heuristic. It is not 
>>a 
>>true measure of correctness. It is not loyal to the truth. "More explanatory" 
>>is 
>>really a heuristic that helps us find hypothesis that are more predictive. 
>>The 
>>true measure of whether a hypothesis is better is simply the most accurate 
>>and 
>>predictive hypothesis. That is the ultimate and true measure of correctness.
>>
>>Also, since we can't measure every possible prediction or every last 
>>prediction 
>>(and we certainly can't predict everything), our measure of predictiveness 
>>can't 
>>possibly be right all the time! We have no choice but to use a heuristic of 
>>some 
>>kind.
>>
>>So, its clear to me that the right hypothesis is "more predictive and then 
>>simpler". But, it is also clear that there will never be a single measure of 
>>this that can be applied to all problems. I hope to eventually find a nice 
>>model 

Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-21 Thread Matt Mahoney
Jim Bromer wrote:
> The fundamental method of Solmonoff Induction is trans-infinite.

The fundamental method is that the probability of a string x is proportional to 
the sum of all programs M that output x weighted by 2^-|M|. That probability is 
dominated by the shortest program, but it is equally uncomputable either way. 
How does this approximation invalidate Solomonoff induction?

Also, please point me to this mathematical community that you claim rejects 
Solomonoff induction. Can you find even one paper that refutes it?

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Wed, July 21, 2010 3:08:13 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


I should have said, It would be unwise to claim that this method could stand as 
an "ideal" for some valid and feasible application of probability.
Jim Bromer


On Wed, Jul 21, 2010 at 2:47 PM, Jim Bromer  wrote:

The fundamental method of Solmonoff Induction is trans-infinite.  Suppose you 
iterate through all possible programs, combining different programs as you go.  
Then you have an infinite number of possible programs which have a 
trans-infinite number of combinations, because each tier of combinations can 
then be recombined to produce a second, third, fourth,... tier of 
recombinations.
> 
>Anyone who claims that this method is the "ideal" for a method of applied 
>probability is unwise.
> Jim Bromer

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-21 Thread Matt Mahoney
Jim Bromer wrote:
> The question was asked whether, given infinite resources could Solmonoff 
>Induction work.  I made the assumption that it was computable and found that 
>it 
>wouldn't work.  

On what infinitely powerful computer did you do your experiment?

> My conclusion suggests, that the use of Solmonoff Induction as an ideal for 
>compression or something like MDL is not only unsubstantiated but based on a 
>massive inability to comprehend the idea of a program that runs every possible 
>program. 

It is sufficient to find the shortest program consistent with past results, not 
all programs. The difference is no more than the language-dependent constant. 
Legg proved this in the paper that Ben and I both pointed you to. Do you 
dispute 
his proof? I guess you don't, because you didn't respond the last 3 times this 
was pointed out to you.

> I am comfortable with the conclusion that the claim that Solomonoff Induction 
>is an "ideal" for compression or induction or anything else is pretty shallow 
>and not based on careful consideration.

I am comfortable with the conclusion that the world is flat because I have a 
gut 
feeling about it and I ignore overwhelming evidence to the contrary.

> There is a chance that I am wrong

So why don't you drop it?

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Tue, July 20, 2010 3:10:40 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


The question was asked whether, given infinite resources could Solmonoff 
Induction work.  I made the assumption that it was computable and found that it 
wouldn't work.  It is not computable, even with infinite resources, for the 
kind 
of thing that was claimed it would do. (I believe that with a governance 
program 
it might actually be programmable) but it could not be used to "predict" (or 
compute the probability of) a subsequent string given some prefix string.  Not 
only is the method impractical it is theoretically inane.  My conclusion 
suggests, that the use of Solmonoff Induction as an ideal for compression or 
something like MDL is not only unsubstantiated but based on a massive inability 
to comprehend the idea of a program that runs every possible program.  

 
I am comfortable with the conclusion that the claim that Solomonoff Induction 
is 
an "ideal" for compression or induction or anything else is pretty shallow and 
not based on careful consideration.
 
There is a chance that I am wrong, but I am confident that there is nothing in 
the definition of Solmonoff Induction that could be used to prove it.
Jim Bromer
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] The Collective Brain

2010-07-21 Thread Matt Mahoney
Mike Tintner wrote:
> The fantasy of a superAGI machine that can grow individually without a vast 
>society supporting it, is another one of the wild fantasies of AGI-ers 
>and Singularitarians that violate truly basic laws of nature. Individual 
>brains 
>cannot flourish individually in the real world, only societies of brains (and 
>bodies) can.

I agree. It is the basis of my AGI design, to supplement a global brain with 
computers. http://mattmahoney.net/agi2.html

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Tue, July 20, 2010 1:50:45 PM
Subject: [agi] The Collective Brain


http://www.ted.com/talks/matt_ridley_when_ideas_have_sex.html?utm_source=newsletter_weekly_2010-07-20&utm_campaign=newsletter_weekly&utm_medium=email

 
Good lecture worth looking at about how trade -  exchange of both goods and 
ideas - has fostered civilisation. Near the end  introduces a v. important idea 
- "the collective brain". In other words, our  apparently individual 
intelligence is actually a collective intelligence. Nobody  he points out 
actually knows how to make a computer mouse, although that may  seem 
counterintuitive  - it's an immensely complex piece of equipment,  simple as it 
may appear, that engages the collective,  interdependent intelligence and 
productive efforts of vast numbers of  people.
 
When you start thinking like that, you realise that  there is v. little we know 
how to do, esp of an intellectual nature,  individually, without the implicit 
and explicit collaboration of vast numbers of  people and sectors of society. 

 
The fantasy of a superAGI machine that can grow  individually without a vast 
society supporting it, is another one of the  wild fantasies of AGI-ers 
and Singularitarians that violate truly  basic laws of nature. Individual 
brains 
cannot flourish individually in the real  world, only societies of brains (and 
bodies) can. 

 
(And of course computers can do absolutely nothing  or in any way survive 
without their human masters - even if it may appear that  way, if you don't 
look 
properly at their whole  operation)
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Of definitions and tests of AGI

2010-07-20 Thread Matt Mahoney
Mike, I think we all agree that we should not have to tell an AGI the steps to 
solving problems. It should learn and figure it out, like the way that people 
figure it out.

The question is how to do that. We know that it is possible. For example, I 
could write a chess program that I could not win against. I could write the 
program in such a way that it learns to improve its game by playing against 
itself or other opponents. I could write it in such a way that initially does 
not know the rules for chess, but instead learns the rules by being given 
examples of legal and illegal moves.

What we have not yet been able to do is scale this type of learning and problem 
solving up to general, human level intelligence. I believe it is possible, but 
it will require lots of training data and lots of computing power. It is not 
something you could do on a PC, and it won't be cheap.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Mon, July 19, 2010 9:07:53 PM
Subject: Re: [agi] Of definitions and tests of AGI


The issue isn't what a computer can do. The issue  is how you structure the 
computer's or any agent's thinking about a  problem. Programs/Turing machines 
are only one way of structuring  thinking/problemsolving - by, among other 
things, giving the  computer a method/process of solution. There is an 
alternative way of  structuring a computer's thinking, which incl., among other 
things, not giving  it a method/ process of solution, but making it rather than 
a human  programmer do the real problemsolving.  More of that another  time.


From: Matt Mahoney 
Sent: Tuesday, July 20, 2010 1:38 AM
To: agi 
Subject: Re: [agi] Of definitions and tests of AGI

Creativity is the good feeling you get when you discover a clever solution  to 
a 
hard problem without knowing the process you used to discover it.

I think a computer could do that.

 -- Matt Mahoney, matmaho...@yahoo.com 





 From: Mike Tintner 
To: agi 
Sent: Mon, July 19, 2010 2:08:28  PM
Subject: Re: [agi] Of  definitions and tests of AGI


Yes that's what people do, but it's not what  programmed computers do.
 
The useful formulation that emerges here  is:
 
narrow AI (and in fact all rational) problems   have *a method of solution*  
(to 
be equated with "general"  method)   - and are programmable (a program is a 
method of  solution)
 
AGI  (and in fact all creative) problems do  NOT have *a method of solution* 
(in 
the general sense)  -  rather  a one.off *way of solving the problem* has to be 
improvised each  time.
 
AGI/creative problems do not in fact have a method  of solution, period. There 
is no (general) method of solving either the toy box  or the build-a-rock-wall 
problem - one essential feature which makes them  AGI.
 
You can learn, as you indicate, from *parts* of any  given AGI/creative 
solution, and apply the lessons to future problems - and  indeed with practice, 
should improve at solving any given kind of AGI/creative  problem. But you can 
never apply a *whole* solution/way to further  problems.
 
P.S. One should add that in terms of computers, we  are talking here of 
*complete, step-by-step* methods of  solution.
 


From: rob levy 
Sent: Monday, July 19, 2010 5:09 PM
To: agi 
Subject: Re: [agi] Of definitions and tests of AGI
  
And are you happy with:
> 
>AGI is about devising *one-off* methods ofproblemsolving (that only apply 
>to 
>the individual problem, and cannot bere-used - at 
>
>
least not in their totality)
> 

Yes exactly, isn't that what people do?  Also, I think that being  able to 
recognize where past solutions can be generalized and where past  solutions can 
be varied and reused is a detail of how intelligence works that is  likely to 
be 
universal.

 
vs
> 
>narrow AI is about applying pre-existing*general* methods of 
>problemsolving  
>(applicable to whole classes ofproblems)?
> 
> 
>
>
>From: rob levy 
>Sent: Monday, July 19, 2010 4:45 PM
>To: agi 
>Subject: Re: [agi] Of definitions and tests ofAGI
>
>Well, solving ANY problem is a little too strong.  This isAGI, not AGH 
>(artificial godhead), though AGH could be an unintendedconsequence ;).  So 
>I 
>would rephrase "solving any problem" as being ableto come up with 
>reasonable 
>approaches and strategies to any problem (just ashumans are able to do).
>
>
>On Mon, Jul 19, 2010 at 11:32 AM, Mike Tintner  
wrote:
>
>Whaddya mean by "solve the problem of how to  solve problems"? Develop a 
>universal approach to solving any problem?  Or find a method of solving a 
>class of problems? Or what?
>>
>>
>>From: rob levy 
>>Sent: Monday, July 19, 2010 1:26 PM
>>To: agi 
>>Subject: Re: [agi] Of definitions and t

Re: [agi] Of definitions and tests of AGI

2010-07-19 Thread Matt Mahoney
Creativity is the good feeling you get when you discover a clever solution to a 
hard problem without knowing the process you used to discover it.

I think a computer could do that.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Mon, July 19, 2010 2:08:28 PM
Subject: Re: [agi] Of definitions and tests of AGI


Yes that's what people do, but it's not what  programmed computers do.
 
The useful formulation that emerges here  is:
 
narrow AI (and in fact all rational) problems   have *a method of solution*  
(to 
be equated with "general"  method)   - and are programmable (a program is a 
method of  solution)
 
AGI  (and in fact all creative) problems do  NOT have *a method of solution* 
(in 
the general sense)  -  rather  a one.off *way of solving the problem* has to be 
improvised each  time.
 
AGI/creative problems do not in fact have a method  of solution, period. There 
is no (general) method of solving either the toy box  or the build-a-rock-wall 
problem - one essential feature which makes them  AGI.
 
You can learn, as you indicate, from *parts* of any  given AGI/creative 
solution, and apply the lessons to future problems - and  indeed with practice, 
should improve at solving any given kind of AGI/creative  problem. But you can 
never apply a *whole* solution/way to further  problems.
 
P.S. One should add that in terms of computers, we  are talking here of 
*complete, step-by-step* methods of  solution.
 


From: rob levy 
Sent: Monday, July 19, 2010 5:09 PM
To: agi 
Subject: Re: [agi] Of definitions and tests of AGI
  
And are you happy with:
> 
>AGI is about devising *one-off* methods ofproblemsolving (that only apply 
>to 
>the individual problem, and cannot bere-used - at 
>
>
least not in their totality)
> 

Yes exactly, isn't that what people do?  Also, I think that being  able to 
recognize where past solutions can be generalized and where past  solutions can 
be varied and reused is a detail of how intelligence works that is  likely to 
be 
universal.

 
vs
> 
>narrow AI is about applying pre-existing*general* methods of 
>problemsolving  
>(applicable to whole classes ofproblems)?
> 
> 
>
>
>From: rob levy 
>Sent: Monday, July 19, 2010 4:45 PM
>To: agi 
>Subject: Re: [agi] Of definitions and tests ofAGI
>
>Well, solving ANY problem is a little too strong.  This isAGI, not AGH 
>(artificial godhead), though AGH could be an unintendedconsequence ;).  So 
>I 
>would rephrase "solving any problem" as being ableto come up with 
>reasonable 
>approaches and strategies to any problem (just ashumans are able to do).
>
>
>On Mon, Jul 19, 2010 at 11:32 AM, Mike Tintner  
wrote:
>
>Whaddya mean by "solve the problem of how to  solve problems"? Develop a 
>universal approach to solving any problem?  Or find a method of solving a 
>class of problems? Or what?
>>
>>
>>From: rob levy 
>>Sent: Monday, July 19, 2010 1:26 PM
>>To: agi 
>>Subject: Re: [agi] Of definitions and tests of  AGI
>>
>>
>> 
>>>However, I see that there are no validdefinitions of AGI that 
>>>explain 
>>>what AGI is generally , and why thesetests are indeed AGI. Google - 
>>>there are v. few defs. of AGI or Strong AI,period.
>
>
>
>
>I like Fogel's idea that intelligence is the ability to "solve the  
>problem 
>of how to solve problems" in new and changing environments.  I  don't 
>think 
>Fogel's method accomplishes this, but the goal he expresses  seems to be 
>the 
>goal of AGI as I understand it. 
>
>
>Rob
>>agi | Archives  | Modify Your Subscription   
>>agi | Archives  | Modify Your Subscription  
>
>agi | Archives  | Modify Your Subscription   
>agi | Archives  | Modify Your Subscription  

agi | Archives  | Modify Your Subscription   
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-18 Thread Matt Mahoney
Jim Bromer wrote:
> The definition of "all possible programs," like the definition of "all 
> possible 
>mathematical functions," is not a proper mathematical problem that can be 
>comprehended in an analytical way.

Finding just the shortest program is close enough because it dominates the 
probability. Or which step in the proof of theorem 1.7.2 
in http://www.vetta.org/documents/disSol.pdf do you disagree with?

You have been saying that you think Solomonoff induction is wrong, but offering 
no argument except your own intuition. So why should we care?

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Sun, July 18, 2010 9:09:36 PM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


Abram,
I was going to drop the discussion, but then I thought I figured out why you 
kept trying to paper over the difference.  Of course, our personal disagreement 
is trivial; it isn't that important.  But the problem with Solomonoff Induction 
is that not only is the output hopelessly tangled and seriously infinite, but 
the input is as well.  The definition of "all possible programs," like the 
definition of "all possible mathematical functions," is not a proper 
mathematical problem that can be comprehended in an analytical way.  I think 
that is the part you haven't totally figured out yet (if you will excuse the 
pun).  "Total program space," does not represent a comprehensible computational 
concept.  When you try find a way to work out feasible computable examples it 
is 
not enough to limit the output string space, you HAVE to limit the program 
space 
in the same way.  That second limitation makes the entire concept of "total 
program space," much too weak for our purposes.  You seem to know this at an 
intuitive operational level, but it seems to me that you haven't truly grasped 
the implications.
 
I say that Solomonoff Induction is computational but I have to use a trick to 
justify that remark.  I think the trick may be acceptable, but I am not sure.  
But the possibility that the concept of "all possible programs," might be 
computational doesn't mean that that it is a sound mathematical concept.  This 
underlies the reason that I intuitively came to the conclusion that Solomonoff 
Induction was transfinite.  However, I wasn't able to prove it because the 
hypothetical concept of "all possible program space," is so pretentious that it 
does not lend itself to mathematical analysis.
 
I just wanted to point this detail out because your implied view that you 
agreed 
with me but "total program space" was "mathematically well-defined" did not 
make 
any sense.
Jim Bromer
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Of definitions and tests of AGI

2010-07-18 Thread Matt Mahoney
http://www.loebner.net/Prizef/loebner-prize.html

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Sun, July 18, 2010 3:10:12 PM
Subject: Re: [agi] Of definitions and tests of AGI

If you can't convince someone, clearly something is wrong with it. I don't 
think 
a "test" is the right way to do this. Which is why I haven't commented much. 
When you understand how to create AGI, it will be obvious that it is AGI or 
that 
it is what you intend it to be. You'll then understand how what you have built 
fits into the bigger scheme of things. There is no such point at which you can 
say something is "AGI" and not "AGI". Intelligence is a very subjective thing 
that really depends on your goals. Someone will always say it is not good 
enough. But if it really works, people will quickly realize it based on results.

What you want is to develop a system that can learn about the world or its 
environment in a general way so that it can solve arbitrary problems, be able 
to 
plan in general ways, act in general ways and perform the types of goals you 
want it to perform. 


Dave


On Sun, Jul 18, 2010 at 3:03 PM, deepakjnath  wrote:

So if I have a system that is close to AGI, I have no way of really knowing it 
right? 

>
>Even if I believe that my system is a true AGI there is no way of convincing 
>the 
>others irrefutably that this system is indeed a AGI not just an advanced AI 
>system.
>
>I have read the toy box problem and rock wall problem, but not many people 
>will 
>still be convinced I am sure.
>
>I wanted to know that if there is any consensus on a general problem which can 
>be solved and only solved by a true AGI. Without such a test bench how will we 
>know if we are moving closer or away from our quest. There is no map.
>
>Deepak
>
>
>
>
>
>On Sun, Jul 18, 2010 at 11:50 PM, Mike Tintner  
wrote:
>
>I realised that what is needed is a *joint*  definition *and*  range of tests 
>of 
>AGI.
>> 
>>Benamin Johnston has submitted one valid test - the  toy box problem. (See 
>>archives).
>> 
>>I have submitted another still simpler valid test -  build a rock wall from 
>>rocks given, (or fill an earth hole with  rocks).
>> 
>>However, I see that there are no valid definitions  of AGI that explain what 
>>AGI 
>>is generally , and why these tests are indeed  AGI. Google - there are v. few 
>>defs. of AGI or Strong AI, period.
>> 
>>The most common: AGI is human-level intelligence -   is an 
>>embarrassing non-starter - what distinguishes human  intelligence? No 
>>explanation offered.
>> 
>>The other two are also inadequate if not as bad:  Ben's "solves a variety of 
>>complex problems in a variety of complex  environments". Nope, so does  a 
>>multitasking narrow AI. Complexity does not  distinguish AGI. Ditto Pei's - 
>>something to do with "insufficient knowledge and  resources..."
>>"Insufficient" is open to narrow AI  interpretations and reducible to 
>>mathematically calculable probabilities.or  uncertainties. That doesn't 
>>distinguish AGI from narrow AI.
>> 
>>The one thing we should all be able to agree on  (but who can be sure?) is 
>that:
>> 
>>** an AGI is a general intelligence system,  capable of independent learning**
>> 
>>i.e. capable of independently learning new  activities/skills with minimal 
>>guidance or even, ideally, with zero guidance (as  humans and animals are) - 
>>and 
>>thus acquiring a "general", "all-round" range of  intelligence..  
>>
>> 
>>This is an essential AGI goal -  the capacity  to keep entering and mastering 
>>new domains of both mental and physical skills  WITHOUT being specially 
>>programmed each time - that crucially distinguishes it  from narrow AI's, 
>>which 
>>have to be individually programmed anew for each new  task. Ben's AGI dog 
>>exemplified this in a v simple way -  the dog is  supposed to be able to 
>>learn 
>>to fetch a ball, with only minimal instructions, as  real dogs do - they can 
>>learn a whole variety of new skills with minimal  instruction.  But I am 
>>confident Ben's dog can't actually do  this.
>> 
>>However, the independent learning def. while  focussing on the distinctive 
>>AGI 
>>goal,  still is not detailed enough by  itself.
>> 
>>It requires further identification of the  **cognitive operations** which 
>>distinguish AGI,  and wh. are exemplified by  the above tests.
>> 
>>[I'll stop there for interruptions/comments &  continue 

Re: [agi] NL parsing

2010-07-16 Thread Matt Mahoney
That that that Buffalo buffalo that Buffalo buffalo buffalo buffalo that 
Buffalo 
buffalo that Buffalo buffalo buffalo.

 -- Matt Mahoney, matmaho...@yahoo.com



- Original Message 
From: Mike Tintner 
To: agi 
Sent: Fri, July 16, 2010 11:05:51 AM
Subject: Re: [agi] NL parsing

Or if you want to be pedantic about caps, the speaker is identifying 3 
buffaloes from Buffalo, & 2 from elsewhere.

Anyone got any other readings?

--
From: "Jiri Jelinek" 
Sent: Friday, July 16, 2010 3:12 PM
To: "agi" 
Subject: [agi] NL parsing

> "Believe it or not, this sentence is grammatically correct and has
> meaning: 'Buffalo buffalo Buffalo buffalo buffalo buffalo Buffalo
> buffalo.'"
>
> source: http://www.mentalfloss.com/blogs/archives/13120
>
> :-)
>
>
> ---
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: 
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com 




---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?&;
Powered by Listbox: http://www.listbox.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] How do we Score Hypotheses?

2010-07-15 Thread Matt Mahoney
Hypotheses are scored using Bayes law. Let D be your observed data and H be 
your 
hypothesis. Then p(H|D) = p(D|H)p(H)/p(D). Since p(D) is constant, you can 
remove it and rank hypotheses by p(D|H)p(H).

p(H) can be estimated using the minimum description length principle or 
Solomonoff induction. Ideally, p(H) = 2^-|H| where |H| is the length (in bits) 
of the description of the hypothesis. The value is language dependent, so this 
method is not perfect.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Thu, July 15, 2010 10:22:44 AM
Subject: Re: [agi] How do we Score Hypotheses?

It is no wonder that I'm having a hard time finding documentation on hypothesis 
scoring. Few can agree on how to do it and there is much debate about it. 


I noticed though that a big reason for the problems is that explanatory 
reasoning is being applied to many diverse problems. I think, like I mentioned 
before, that people should not try to come up with a single universal rule set 
for applying explanatory reasoning to every possible problem. So, maybe that's 
where the hold up is. 


I've been testing my ideas out on complex examples. But now I'm going to go 
back 
to simplified model testing (although not as simple as black squares :) ) and 
work my way up again. 


Dave


On Wed, Jul 14, 2010 at 12:59 PM, David Jones  wrote:

Actually, I just realized that there is a way to included inductive knowledge 
and experience into this algorithm. Inductive knowledge and experience about a 
specific object or object type can be exploited to know which hypotheses in the 
past were successful, and therefore which hypothesis is most likely. By 
choosing 
the most likely hypothesis first, we skip a lot of messy hypothesis comparison 
processing and analysis. If we choose the right hypothesis first, all we really 
have to do is verify that this hypothesis reveals in the data what we expect to 
be there. If we confirm what we expect, that is reason enough not to look for 
other hypotheses because the data is explained by what we originally believed 
to 
be likely. We only look for additional hypotheses when we find something 
unexplained. And even then, we don't look at the whole problem. We only look at 
what we have to to explain the unexplained data. In fact, we could even ignore 
the unexplained data if we believe, from experience, that it isn't pertinent. 

>
>I discovered this because I'm analyzing how a series of hypotheses are 
>navigated 
>when analyzing images. It seems to me that it is done very similarly to way we 
>do it. We sort of confirm what we expect and try to explain what we don't 
>expect. We try out hypotheses in a sort of trial and error manor and see how 
>each hypothesis affects what we find in the image. If we confirm things 
>because 
>of the hypothesis, we are likely to keep it. We keep going, navigating the 
>tree 
>of hypotheses, conflicts and unexpected observations until we find a good 
>hypothesis. Something like that. I'm attempting to construct an algorithm for 
>doing this as I analyze specific problems. 
>
>
>Dave
>
>
>
>On Wed, Jul 14, 2010 at 10:22 AM, David Jones  wrote:
>
>What do you mean by definitive events? 
>>
>>I guess the first problem I see with my approach is that the movement of the 
>>window is also a hypothesis. I need to analyze it in more detail and see how 
>>the 
>>tree of hypotheses affects the hypotheses regarding the "e"s on the windows. 
>>
>>
>>What I believe is that these problems can be broken down into types of 
>>hypotheses,  types of events and types of relationships. then those types can 
>>be 
>>reasoned about in a general way. If possible, then you have a method for 
>>reasoning about any object that is covered by the types of hypotheses, events 
>>and relationships that you have defined.
>>
>>How to reason about specific objects should not be preprogrammed. But, I 
>>think 
>>the solution to this part of AGI is to find general ways to reason about a 
>>small 
>>set of concepts that can be combined to describe specific objects and 
>>situations. 
>>
>>
>>There are other parts to AGI that I am not considering yet. I believe the 
>>problem has to be broken down into separate pieces and understood before 
>>putting 
>>it back together into a complete system. I have not covered inductive 
>>learning 
>>for example, which would be an important part of AGI. I have also not yet 
>>incorporated learned experience into the algorithm, which is also important. 
>>
>>
>>The general AI problem is way too complicated to consider all at once. I 
>>simply 
>>can't solve hypothesis generation, co

Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-15 Thread Matt Mahoney
Jim Bromer wrote:
> Since you cannot fully compute every string that may be produced at a certain 
>iteration, you cannot make the claim that you even know the probabilities of 
>any 
>possible string before infinity and therefore your claim that the sum of the 
>probabilities can be computed is not provable.
> 
> But I could be wrong.

Could be. Theorem 1.7.2 in http://www.vetta.org/documents/disSol.pdf proves 
that 
finding just the shortest program that outputs x gives you a probability for x 
close to the result you would get if you found all of the (infinite number of) 
programs that output x. Either number could be used for Solomonoff induction 
because the difference is bounded only by the choice of language.
 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Thu, July 15, 2010 8:18:13 AM
Subject: Re: [agi] Comments On My Skepticism of Solomonoff Induction


On Wed, Jul 14, 2010 at 7:46 PM, Abram Demski  wrote:

Jim,
>
>There is a simple proof of convergence for the sum involved in defining the 
>probability of a given string in the Solomonoff distribution:
>
>At its greatest, a particular string would be output by *all* programs. In 
>this 
>case, its sum would come to 1. This puts an upper bound on the sum. Since 
>there 
>is no subtraction, there is a lower bound at 0 and the sum monotonically 
>increases as we take the limit. Knowing these facts, suppose it *didn't* 
>converge. It must then increase without bound, since it cannot fluctuate back 
>and forth (it can only go up). But this contradicts the upper bound of 1. So, 
>the sum must stop at 1 or below (and in fact we can prove it stops below 1, 
>though we can't say where precisely without the infinite computing power 
>required to compute the limit).
>
>--Abram
 
I believe that Solomonoff Induction would be computable given infinite time and 
infinite resources (the Godel Theorem fits into this category) but some people 
disagree for reasons I do not understand.  

 
If it is not computable then it is not a mathematical theorem and the question 
of whether the sum of probabilities equals 1 is pure fantasy.
 
If it is computable then the central issue is whether it could (given infinite 
time and infinite resources) be used to determine the probability of a 
particular string being produced from all possible programs.  The question 
about 
the sum of all the probabilities is certainly an interesting question. However, 
the problem of making sure that the function was actually computable would 
interfere with this process of determining the probability of each particular 
string that can be produced.  For example, since some strings would be 
infinite, 
the computability problem makes it imperative that the infinite strings be 
partially computed at an iteration (or else the function would be hung up at 
some particular iteration and the infinite other calculations could not be 
considered computable).  

 
My criticism is that even though I believe the function may be theoretically 
computable, the fact that each particular probability (of each particular 
string 
that is produced) cannot be proven to approach a limit through mathematical 
analysis, and since the individual probabilities will fluctuate with each new 
string that is produced, one would have to know how to reorder the production 
of 
the probabilities in order to demonstrate that the individual probabilities do 
approach a limit.  If they don't, then the claim that this function could be 
used to define the probabilities of a particular string from all possible 
program is unprovable.  (Some infinite calculations fluctuate infinitely.)  
Since you do not have any way to determine how to reorder the infinite 
probabilities a priori, your algorithm would have to be able to compute all 
possible reorderings to find the ordering and filtering of the computations 
that 
would produce evaluable limits.  Since there are trans infinite rearrangements 
of an infinite list (I am not sure that I am using the term 'trans infinite' 
properly) this shows that the conclusion that the theorem can be used to derive 
the desired probabilities is unprovable through a variation of Cantor's 
Diagonal 
Argument, and that you can't use Solomonoff Induction the way you have been 
talking about using it.
 
Since you cannot fully compute every string that may be produced at a certain 
iteration, you cannot make the claim that you even know the probabilities of 
any 
possible string before infinity and therefore your claim that the sum of the 
probabilities can be computed is not provable.
 
But I could be wrong.
Jim Bromer
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify 

Re: [agi] What is the smallest set of operations that can potentially define everything and how do you combine them ?

2010-07-14 Thread Matt Mahoney
Michael Swan wrote:
> What 3456/6 ?
> we don't know, at least not from the top of our head.

No, it took me about 10 or 20 seconds to get 576. Starting with the first 
digit, 
3/6 = 1/2 (from long term memory) and 3 is in the thousands place, so 1/2 of 
1000 is 500 (1/2 = .5 from LTM). I write 500 into short term memory (STM), 
which 
only has enough space to hold about 7 digits. Then to divide 45/6 I get 42/6 = 
7 
with a remainder of 3, or 7.5, but since this is in the tens place I get 75. I 
put 75 in STM, add to 500 to get 575, put the result back in STM replacing 500 
and 75 for which there is no longer room. Finally, 6/6 = 1, which I add to 575 
to get 576. I hold this number in STM long enough to check with a calculator.

One could argue that this calculation in my head uses a loop iterator (in STM) 
to keep track of which digit I am working on. It definitely involves a sequence 
of instructions with intermediate results being stored temporarily. The brain 
can only execute 2 or 3 sequential instructions per second and has very limited 
short term memory, so it needs to draw from a large database of rules to 
perform 
calculations like this. A calculator, being faster and having more RAM, is able 
to use simpler but more tedious algorithms such as converting to binary, 
division by shift and subtract, and converting back to decimal. Doing this with 
a carbon based computer would require pencil and paper to make up for lack of 
STM, and it would require enough steps to have a high probability of making a 
mistake.

Intelligence = knowledge + computing power. The human brain has a lot of 
knowledge. The calculator has less knowledge, but makes up for it in speed and 
memory.

 -- Matt Mahoney, matmaho...@yahoo.com



- Original Message 
From: Michael Swan 
To: agi 
Sent: Wed, July 14, 2010 7:53:33 PM
Subject: Re: [agi] What is the smallest set of operations that can potentially  
define everything and how do you combine them ?

On Wed, 2010-07-14 at 07:48 -0700, Matt Mahoney wrote:
> Actually, Fibonacci numbers can be computed without loops or recursion.
> 
> int fib(int x) {
>   return round(pow((1+sqrt(5))/2, x)/sqrt(5));
> }
;) I know. I was wondering if someone would pick up on it. This won't
prove that brains have loops though, so I wasn't concerned about the
shortcuts. 
> unless you argue that loops are needed to compute sqrt() and pow().
> 
I would find it extremely unlikely that brains have *, /, and even more
unlikely to have sqrt and pow inbuilt. Even more unlikely, even if it
did have them, to figure out how to combine them to round(pow((1
+sqrt(5))/2, x)/sqrt(5)). 

Does this mean we should discount all maths that use any complex
operations ? 

I suspect the brain is full of look-up tables mainly, with some fairly
primitive methods of combining the data. 

eg What's 6 / 3 ?
ans = 2 most people would get that because it's been wrote learnt, a
common problem.

What 3456/6 ?
we don't know, at least not from the top of our head.


> The brain and DNA use redundancy and parallelism and don't use loops because 
> their operations are slow and unreliable. This is not necessarily the best 
> strategy for computers because computers are fast and reliable but don't have 
> a 
>
> lot of parallelism.

The brains "slow and unreliable" methods I think are the price paid for
generality and innately unreliable hardware. Imagine writing a computer
program that runs for 120 years without crashing and surviving damage
like a brain can. I suspect the perfect AGI program is a rigorous
combination of the 2. 


> 
>  -- Matt Mahoney, matmaho...@yahoo.com
> 
> 
> 
> - Original Message 
> From: Michael Swan 
> To: agi 
> Sent: Wed, July 14, 2010 12:18:40 AM
> Subject: Re: [agi] What is the smallest set of operations that can 
> potentially  
>
> define everything and how do you combine them ?
> 
> Brain loops:
> 
> 
> Premise:
> Biological brain code does not contain looping constructs, or the
> ability to creating looping code, (due to the fact they are extremely
> dangerous on unreliable hardware) except for 1 global loop that fires
> about 200 times a second.
> 
> Hypothesis:
> Brains cannot calculate "iterative" problems quickly, where calculations
> in the previous iteration are needed for the next iteration and, where
> brute force operations are the only valid option.
> 
> Proof:
> Take as an example, Fibonacci numbers
> http://en.wikipedia.org/wiki/Fibonacci_number
> 
> What are the first 100 Fibonacci numbers?
> 
> int Fibonacci[102];
> Fibonacci[0] = 0;
> Fibonacci[1] = 1;
> for(int i = 0; i < 100; i++)
> {
> // Getting the next Fibonacci number relies on the previous values
> Fibonacci[i+2] = Fibonacci[i] + Fibonacci[i+1];
> }  
> 
> My brai

Re: [agi] Comments On My Skepticism of Solomonoff Induction

2010-07-14 Thread Matt Mahoney
Jim Bromer wrote:
> Last week I came up with a sketch that I felt showed that Solomonoff 
> Induction 
>was incomputable in practice using a variation of Cantor's Diagonal Argument.

Cantor proved that there are more sequences (infinite length strings) than 
there 
are (finite length) strings, even though both sets are infinite. This means 
that 
some, but not all, sequences have finite length descriptions or are the output 
of finite length programs (which is the same thing in a more formal sense). For 
example, the digits of pi or sqrt(2) are infinite length sequences that have 
finite descriptions (or finite programs that output them). There are many more 
sequences that don't have finite length descriptions, but unfortunately I can't 
describe any of them except to say they contain infinite amounts of random data.

Cantor does not prove that Solomonoff induction is not computable. That was 
proved by Kolmogorov (and also by Solomonoff). Solomonoff induction says to use 
the shortest program that outputs the observed sequence to predict the next 
symbol. However, there is no procedure for finding the length of the shortest 
description. The proof sketch is that if there were, then I could describe "the 
first string that cannot be described in less than a million bits" even though 
I 
just did. The formal proof 
is 
http://en.wikipedia.org/wiki/Kolmogorov_complexity#Incomputability_of_Kolmogorov_complexity


I think your confusion is using the uncomputability of Solomonoff induction to 
question its applicability. That is an experimental question, not one of 
mathematics. The validity of using the shortest or simplest explanation of the 
past to predict the future was first observed by William of Ockham in the 
1400's. It is standard practice in all fields of science. The minimum 
description length principle is applicable to all branches of machine learning.

However, in the conclusion 
of http://mattmahoney.net/dc/dce.html#Section_Conclusion I argue for Solomonoff 
induction on the basis of physics. Solomonoff induction supposes that all 
observable strings are finite prefixes of computable sequences. Occam's Razor 
might not hold if it were possible for the universe to produce uncomputable 
sequences, i.e. infinite sources of random data. I argue that is not possible 
because the observable universe if finitely computable according to the laws of 
physics as they are now understood.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Wed, July 14, 2010 11:29:13 AM
Subject: [agi] Comments On My Skepticism of Solomonoff Induction


Last week I came up with a sketch that I felt showed that Solomonoff Induction 
was incomputable in practice using a variation of Cantor's Diagonal Argument.  
I 
wondered if my argument made sense or not.  I will explain why I think it did.
 
First of all, I should have started out by saying something like, Suppose 
Solomonoff Induction was computable, since there is some reason why people feel 
that it isn't.
 
Secondly I don't think I needed to use Cantor's Diagonal Argument (for the in 
practice case), because it would be sufficient to point out that since it was 
impossible to say whether or not the probabilities ever approached any 
sustained 
(collared) limits due to the lack of adequate mathematical definition of the 
concept "all programs", it would be impossible to make the claim that they were 
actual representations of the probabilities of all programs that could produce 
certain strings.
 
But before I start to explain why I think my variation of the Diagonal Argument 
was valid, I would like to make another comment about what was being claimed.
 
Take a look at the n-ary expansion of the square root of 2 (such as the decimal 
expansion or the binary expansion).  The decimal expansion or the binary 
expansion of the square root of 2 is an infinite string.  To say that the 
algorithm that produces the value is "predicting" the value is a torturous use 
of meaning of the word 'prediction'.  Now I have less than perfect grammar, but 
the idea of prediction is so important in the field of intelligence that I do 
not feel that this kind of reduction of the concept of prediction is 
illuminating.  

 
Incidentally, There are infinite ways to produce the square root of 2 (sqrt 2 
+1-1, sqrt2 +2-2, sqrt2 +3-3,...).  So the idea that the square root of 2 is 
unlikely is another stretch of conventional thinking.  But since there are an 
infinite ways for a program to produce any number (that can be produced by a 
program) we would imagine that the probability that one of the infinite ways to 
produce the square root of 2 approaches 0 but never reaches it.  We can imagine 
it, but we cannot prove that this occurs in Solomonoff Induction because 
Solomonoff Induction is not limited to just this class of programs (which 

Re: [agi] What is the smallest set of operations that can potentially define everything and how do you combine them ?

2010-07-14 Thread Matt Mahoney
Actually, Fibonacci numbers can be computed without loops or recursion.

int fib(int x) {
  return round(pow((1+sqrt(5))/2, x)/sqrt(5));
}

unless you argue that loops are needed to compute sqrt() and pow().

The brain and DNA use redundancy and parallelism and don't use loops because 
their operations are slow and unreliable. This is not necessarily the best 
strategy for computers because computers are fast and reliable but don't have a 
lot of parallelism.

 -- Matt Mahoney, matmaho...@yahoo.com



- Original Message 
From: Michael Swan 
To: agi 
Sent: Wed, July 14, 2010 12:18:40 AM
Subject: Re: [agi] What is the smallest set of operations that can potentially  
define everything and how do you combine them ?

Brain loops:


Premise:
Biological brain code does not contain looping constructs, or the
ability to creating looping code, (due to the fact they are extremely
dangerous on unreliable hardware) except for 1 global loop that fires
about 200 times a second.

Hypothesis:
Brains cannot calculate "iterative" problems quickly, where calculations
in the previous iteration are needed for the next iteration and, where
brute force operations are the only valid option.

Proof:
Take as an example, Fibonacci numbers
http://en.wikipedia.org/wiki/Fibonacci_number

What are the first 100 Fibonacci numbers?

int Fibonacci[102];
Fibonacci[0] = 0;
Fibonacci[1] = 1;
for(int i = 0; i < 100; i++)
{
// Getting the next Fibonacci number relies on the previous values
Fibonacci[i+2] = Fibonacci[i] + Fibonacci[i+1];
}  

My brain knows the process to solve this problem but it can't directly
write a looping construct into itself. And so it solves it very slowly
compared to a computer. 

The brain probably consists of vast repeating look-up tables. Of course,
run in parallel these seem fast.


DNA has vast tracks of repeating data. Why would DNA contain repeating
data, instead of just having the data once and the number of times it's
repeated like in a loop? One explanation is that DNA can't do looping
construct either.



On Wed, 2010-07-14 at 02:43 +0100, Mike Tintner wrote:
> Michael: We can't do operations that
> require 1,000,000 loop iterations.  I wish someone would give me a PHD
> for discovering this ;) It far better describes our differences than any
> other theory.
> 
> Michael,
> 
> This isn't a competitive point - but I think I've made that point several 
> times (and so of course has Hawkins). Quite obviously, (unless you think the 
> brain has fabulous hidden powers), it conducts searches and other operations 
> with extremely few limited steps, and nothing remotely like the routine 
> millions to billions of current computers.  It must therefore work v. 
> fundamentally differently.
> 
> Are you saying anything significantly different to that?
> 
> --
> From: "Michael Swan" 
> Sent: Wednesday, July 14, 2010 1:34 AM
> To: "agi" 
> Subject: Re: [agi] What is the smallest set of operations that can 
> potentially  define everything and how do you combine them ?
> 
> >
> > On Tue, 2010-07-13 at 07:00 -0400, Ben Goertzel wrote:
> >> Well, if you want a simple but complete operator set, you can go with
> >>
> >> -- Schonfinkel combinator plus two parentheses
> >>
> > I'll check this out soon.
> >> or
> >>
> >> -- S and K combinator plus two parentheses
> >>
> >> and I suppose you could add
> >>
> >> -- input
> >> -- output
> >> -- forget
> >>
> >> statements to this, but I'm not sure what this gets you...
> >>
> >> Actually, adding other operators doesn't necessarily
> >> increase the search space your AI faces -- rather, it
> >> **decreases** the search space **if** you choose the right operators, 
> >> that
> >> encapsulate regularities in the environment faced by the AI
> >
> > Unfortunately, an AGI needs to be absolutely general. You are right that
> > higher level concepts reduce combinations, however, using them, will
> > increase combinations for "simpler" operator combinations, and if you
> > miss a necessary operator, then some concepts will be impossible to
> > achieve. The smallest set can define higher level concepts, these
> > concepts can be later integrated as "single" operations, which means
> > using "operators than can be understood in terms of smaller operators"
> > in the beginning, will definitely increase you combinations later on.
> >
> > The smallest operator set is like absolute zero. It has a defined end. A
> > defined way of finding out what they are.
> >
&g

Re: [agi] Mechanical Analogy for Neural Operation!

2010-07-12 Thread Matt Mahoney
Steve Richfield wrote:
> No, I am NOT proposing building mechanical contraptions, just using the 
> concept 
>to compute neuronal characteristics (or AGI formulas for learning).

Funny you should mention that. Ross Ashby actually built such a device in 1948 
called a homeostat ( http://en.wikipedia.org/wiki/Homeostat ), a fully 
interconnected neural network with 4 neurons using mechanical components and 
vacuum tubes. Synaptic weights were implemented by motor driven water filled 
potentiometers in which electrodes moved through a tank to vary the electrical 
resistance. It implemented a type of learning algorithm in which weights were 
varied using a rotating switch wired randomly using the RAND book of a million 
random digits. He described the device in his 1960 book, Design for a Brain.
 -- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Mon, July 12, 2010 2:02:20 AM
Subject: [agi] Mechanical Analogy for Neural Operation!

Everyone has heard about the water analogy for electrical operation. I have a 
mechanical analogy for neural operation that just might be "solid" enough to 
compute at least some characteristics optimally.

No, I am NOT proposing building mechanical contraptions, just using the concept 
to compute neuronal characteristics (or AGI formulas for learning).

Suppose neurons were mechanical contraptions, that receive inputs and 
communicate outputs via mechanical movements. If one or more of the neurons 
connected to an output of a neuron, can't make sense of a given input given its 
other inputs, then its mechanism would physically resist the several inputs 
that 
didn't make mutual sense because its mechanism would jam, with the resistance 
possibly coming from some downstream neuron.

This would utilize position to resolve opposing forces, e.g. one "force" being 
the observed inputs, and the other "force" being that they don't make sense, 
suggest some painful outcome, etc. In short, this would enforce the sort of 
equation over the present formulaic view of neurons (and AGI coding) that I 
have 
suggested in past postings may be present, and show that the math may not be 
all 
that challenging.

Uncertainty would be expressed in stiffness/flexibility, computed limitations 
would be handled with over-running clutches, etc.

Propagation of forces would come close (perfect?) to being able to identify 
just 
where in a complex network something should change to learn as efficiently as 
possible.

Once the force concentrates at some point, it then "gives", something slips or 
bends, to unjam the mechanism. Thus, learning is effected.

Note that this suggests little difference between forward propagation and 
backwards propagation, though real-world wet design considerations would 
clearly 
prefer fast mechanisms for forward propagation, and compact mechanisms for 
backwards propagation.

Epiphany or mania?

Any thoughts?

Steve

 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Solomonoff Induction is Not "Universal" and Probability is not "Prediction"

2010-07-09 Thread Matt Mahoney
Ben Goertzel wrote:
>> Secondly, since it cannot be computed it is useless.  Third, it is not the 
>> sort 
>>of thing that is useful for AGI in the first place.

> I agree with these two statements

The principle of Solomonoff induction can be applied to computable subsets of 
the (infinite) hypothesis space. For example, if you are using a neural network 
to make predictions, the principle says to use the smallest network that 
computes the past training data.
 -- Matt Mahoney, matmaho...@yahoo.com





From: Ben Goertzel 
To: agi 
Sent: Fri, July 9, 2010 7:56:53 AM
Subject: Re: [agi] Solomonoff Induction is Not "Universal" and Probability is 
not "Prediction"




On Fri, Jul 9, 2010 at 7:49 AM, Jim Bromer  wrote:

Abram,
>Solomoff Induction would produce poor "predictions" if it could be used to 
>compute them.  
>

Solomonoff induction is a mathematical, not verbal, construct.  Based on the 
most obvious mapping from the verbal terms you've used above into mathematical 
definitions in terms of which Solomonoff induction is constructed, the above 
statement of yours is FALSE.

If you're going to argue against a mathematical theorem, your argument must be 
mathematical not verbal.  Please explain one of

1) which step in the proof about Solomonoff induction's effectiveness you 
believe is in error

2) which of the assumptions of this proof you think is inapplicable to real 
intelligence [apart from the assumption of infinite or massive compute 
resources]

Otherwise, your statement is in the same category as the statement by the 
protagonist of Dostoesvky's "Notes from the Underground" --

"I admit that two times two makes four is an excellent thing, but if we are to 
give everything its due, two times two makes five is sometimes a very charming 
thing too."

;-)

 
Secondly, since it cannot be computed it is useless.  Third, it is not the sort 
of thing that is useful for AGI in the first place.

I agree with these two statements

-- ben G 


agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Solomonoff Induction is Not "Universal" and Probability is not "Prediction"

2010-07-07 Thread Matt Mahoney
Who is talking about efficiency? An infinite sequence of uncomputable values is 
still just as uncomputable. I don't disagree that AIXI and Solomonoff induction 
are not computable. But you are also arguing that they are wrong.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Wed, July 7, 2010 6:40:52 PM
Subject: Re: [agi] Solomonoff Induction is Not "Universal" and Probability is 
not "Prediction"


Matt,
But you are still saying that Solomonoff Induction has to be recomputed for 
each 
possible combination of bit value aren't you?  Although this doesn't matter 
when 
you are dealing with infinite computations in the first place, it does matter 
when you are wondering if this has anything to do with AGI and compression 
efficiencies.
Jim Bromer

On Wed, Jul 7, 2010 at 5:44 PM, Matt Mahoney  wrote:

 Jim Bromer wrote:
>> But, a more interesting question is, given that the first digits are 000, 
>> what 
>>are the chances that the next digit will be 1?  Dim Induction will report .5, 
>>which of course is nonsense and a whole less useful than making a rough guess.
>
>
>Wrong. The probability of a 1 is p(0001)/(p()+p(0001)) where the 
>probabilities are computed using Solomonoff induction. A program that outputs 
> will be shorter in most languages than a program that outputs 0001, so 0 
>is 
>the most likely next bit.
>
>
>More generally, probability and prediction are equivalent by the chain rule. 
>Given any 2 strings x followed by y, the prediction p(y|x) = p(xy)/p(x).
>
> -- Matt Mahoney, matmaho...@yahoo.com 
>
>
>
>
>

 From: Jim Bromer 
>To: agi 
>Sent: Wed, July 7, 2010 10:10:37 AM
>Subject: [agi] Solomonoff Induction is Not "Universal" and Probability is not 
>"Prediction"
> 
>
>
>Suppose you have sets of "programs" that produce two strings.  One set of 
>outputs is 00 and the other is 11. Now suppose you used these sets of 
>programs to chart the probabilities of the output of the strings.  If the two 
>strings were each output by the same number of programs then you'd have a .5 
>probability that either string would be output.  That's ok.  But, a more 
>interesting question is, given that the first digits are 000, what are the 
>chances that the next digit will be 1?  Dim Induction will report .5, which of 
>course is nonsense and a whole less useful than making a rough guess.
> 
>But, of course, Solomonoff Induction purports to be able, if it was feasible, 
>to 
>compute the possibilities for all possible programs.  Ok, but now, try 
>thinking 
>about this a little bit.  If you have ever tried writing random program 
>instructions what do you usually get?  Well, I'll take a hazard and guess (a 
>lot 
>better than the bogus method of confusing shallow probability with 
>"prediction" 
>in my example) and say that you will get a lot of programs that crash.  Well, 
>most of my experiments with that have ended up with programs that go into an 
>infinite loop or which crash.  Now on a universal Turing machine, the results 
>would probably look a little different.  Some strings will output nothing and 
>go 
>into an infinite loop.  Some programs will output something and then either 
>stop 
>outputting anything or start outputting an infinite loop of the same 
>substring.  
>Other programs will go on to infinity producing something that looks like 
>random 
>strings.  But the idea that all possible programs would produce well 
>distributed 
>strings is complete hogwash.  Since Solomonoff Induction does not define what 
>kind of programs should be used, the assumption that the distribution would 
>produce useful data is absurd.  In particular, the use of the method to 
>determine the probability based given an initial string (as in what follows 
>given the first digits are 000) is wrong as in really wrong.  The idea that 
>this 
>crude probability can be used as "prediction" is unsophisticated.
> 
>Of course you could develop an infinite set of Solomonoff Induction values for 
>each possible given initial sequence of digits.  Hey when you're working with 
>infeasible functions why not dream anything?
> 
>I might be wrong of course.  Maybe there is something you guys haven't been 
>able 
>to get across to me.  Even if you can think for yourself you can still make 
>mistakes.  So if anyone has actually tried writing a program to output all 
>possible programs (up to some feasible point) on a Turing Machine simulator, 
>let 
>me know how it went.
> 
>Jim Bromer
> 
>agi | Archives  | Modify Your Subscription  
>agi | Archiv

Re: [agi] Hutter - A fundamental misdirection?

2010-07-07 Thread Matt Mahoney
Gorrell and Webb describe a neural implementation of LSA that seems more 
biologically plausible than the usual matrix factoring implementation.
http://www.dcs.shef.ac.uk/~genevieve/gorrell_webb.pdf
 
In the usual implementation, a word-word matrix A is factored to A = USV where 
S 
is diagonal (containing eigenvalues), and then the smaller elements of S are 
discarded. In the Gorrell model, U and V are the weights of a 3 layer neural 
network mapping words to words, and the nonzero elements of S represent the 
semantic space in the middle layer. As the network is trained, neurons are 
added 
to S. Thus, the network is trained online in a single pass, unlike factoring, 
which is offline.

-- Matt Mahoney, matmaho...@yahoo.com





From: Gabriel Recchia 
To: agi 
Sent: Wed, July 7, 2010 12:12:00 PM
Subject: Re: [agi] Hutter - A fundamental misdirection?

> In short, instead of a "pot of neurons", we might instead have a pot of 
> dozens 
>of types of 
>
> neurons that each have their own complex rules regarding what other types of 
>neurons they 
>
> can connect to, and how they process information...

> ...there is plenty of evidence (from the slowness of evolution, the large 
>number (~200) 
>
> of neuron types, etc.), that it is many-layered and quite complex...

The disconnect between the low-level neural hardware and the implementation of 
algorithms that build conceptual spaces via dimensionality reduction--which 
generally ignore facts such as the existence of different types of neurons, the 
apparently hierarchical organization of neocortex, etc.--seems significant. 
Have 
there been attempts to develop computational models capable of LSA-style feats 
(e.g., constructing a vector space in which words with similar meanings tend to 
be relatively close to each other) that take into account basic facts about how 
neurons actually operate (ideally in a more sophisticated way than the nodes of 
early connectionist networks which, as we now know, are not particularly 
neuron-like at all)? If so, I would love to know about them.



On Tue, Jun 29, 2010 at 3:02 PM, Ian Parker  wrote:

The paper seems very similar in principle to LSA. What you need for a concept 
vector  (or position) is the application of LSA followed by K-Means which will 
give you your concept clusters.
>
>
>I would not knock Hutter too much. After all LSA reduces {primavera, 
>mamanthal, 
>salsa, resorte} to one word giving 2 bits saving on Hutter.
>
>
>
>
>  - Ian Parker
>
>
>
>On 29 June 2010 07:32, rob levy  wrote:
>
>Sorry, the link I included was invalid, this is what I meant: 
>>
>>
>>http://www.geog.ucsb.edu/~raubal/Publications/RefConferences/ICSC_2009_AdamsRaubal_Camera-FINAL.pdf
>>
>>
>>
>>
>>On Tue, Jun 29, 2010 at 2:28 AM, rob levy  wrote:
>>
>>On Mon, Jun 28, 2010 at 5:23 PM, Steve Richfield  
>>wrote:
>>>
>>>Rob,
>>>>
>>>>I just LOVE opaque postings, because they identify people who see things 
>>>>differently than I do. I'm not sure what you are saying here, so I'll make 
>>>>some 
>>>>"random" responses to exhibit my ignorance and elicit more explanation.
>>>>
>>>>
>>
>>
>>>I think based on what you wrote, you understood (mostly) what I was trying 
>>>to 
>>>get across.  So I'm glad it was at least quasi-intelligible. :)
>>> 
>>> It sounds like this is a finer measure than the "dimensionality" that I was 
>>>referencing. However, I don't see how to reduce anything as quantized as 
>>>dimensionality into finer measures. Can you say some more about this?
>>>>
>>>>
>>
>>
>>>I was just referencing Gardenfors' research program of "conceptual spaces" 
>>>(I 
>>>was intentionally vague about committing to this fully though because I 
>>>don't 
>>>necessarily think this is the whole answer).  Page 2 of this article 
>>>summarizes 
>>>it pretty succinctly: 
>>>http://www.geog.ucsb.edu/.../ICSC_2009_AdamsRaubal_Camera-FINAL.pdf
>>>
>>>
>>> 
>>>However, different people's brains, even the brains of identical twins, have 
>>>DIFFERENT mappings. This would seem to mandate experience-formed topology.
>>>> 
>>>>
>>
>>
>>>Yes definitely.
>>> 
>>>Since these conceptual spaces that structure sensorimotor 
>>>expectation/prediction 
>>>(including in higher order embodied exploration of concepts I think) are 
>>>multidimensional spaces, it seems likely that some kind of neur

Re: [agi] Solomonoff Induction is Not "Universal" and Probability is not "Prediction"

2010-07-07 Thread Matt Mahoney
 Jim Bromer wrote:
> But, a more interesting question is, given that the first digits are 000, 
> what 
>are the chances that the next digit will be 1?  Dim Induction will report .5, 
>which of course is nonsense and a whole less useful than making a rough guess.

Wrong. The probability of a 1 is p(0001)/(p()+p(0001)) where the 
probabilities are computed using Solomonoff induction. A program that outputs 
 will be shorter in most languages than a program that outputs 0001, so 0 
is 
the most likely next bit.

More generally, probability and prediction are equivalent by the chain rule. 
Given any 2 strings x followed by y, the prediction p(y|x) = p(xy)/p(x).

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Wed, July 7, 2010 10:10:37 AM
Subject: [agi] Solomonoff Induction is Not "Universal" and Probability is not 
"Prediction"


Suppose you have sets of "programs" that produce two strings.  One set of 
outputs is 00 and the other is 11. Now suppose you used these sets of 
programs to chart the probabilities of the output of the strings.  If the two 
strings were each output by the same number of programs then you'd have a .5 
probability that either string would be output.  That's ok.  But, a more 
interesting question is, given that the first digits are 000, what are the 
chances that the next digit will be 1?  Dim Induction will report .5, which of 
course is nonsense and a whole less useful than making a rough guess.
 
But, of course, Solomonoff Induction purports to be able, if it was feasible, 
to 
compute the possibilities for all possible programs.  Ok, but now, try thinking 
about this a little bit.  If you have ever tried writing random program 
instructions what do you usually get?  Well, I'll take a hazard and guess (a 
lot 
better than the bogus method of confusing shallow probability with "prediction" 
in my example) and say that you will get a lot of programs that crash.  Well, 
most of my experiments with that have ended up with programs that go into an 
infinite loop or which crash.  Now on a universal Turing machine, the results 
would probably look a little different.  Some strings will output nothing and 
go 
into an infinite loop.  Some programs will output something and then either 
stop 
outputting anything or start outputting an infinite loop of the same substring. 
 
Other programs will go on to infinity producing something that looks like 
random 
strings.  But the idea that all possible programs would produce well 
distributed 
strings is complete hogwash.  Since Solomonoff Induction does not define what 
kind of programs should be used, the assumption that the distribution would 
produce useful data is absurd.  In particular, the use of the method to 
determine the probability based given an initial string (as in what follows 
given the first digits are 000) is wrong as in really wrong.  The idea that 
this 
crude probability can be used as "prediction" is unsophisticated.
 
Of course you could develop an infinite set of Solomonoff Induction values for 
each possible given initial sequence of digits.  Hey when you're working with 
infeasible functions why not dream anything?
 
I might be wrong of course.  Maybe there is something you guys haven't been 
able 
to get across to me.  Even if you can think for yourself you can still make 
mistakes.  So if anyone has actually tried writing a program to output all 
possible programs (up to some feasible point) on a Turing Machine simulator, 
let 
me know how it went.
 
Jim Bromer
 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-04 Thread Matt Mahoney
Jim Bromer wrote:
>  However, I don't accept that it is feasible to make those calculations since 
> an examination of the infinite programs that could output each individual 
> string would be required.

In fact, finding just the most likely program is not computable. Minimum 
description length (MDL), Solomonoff induction and Occam's Razor do have 
practical implications because they can be used to choose among a restricted 
but computable set of hypothesis. For example, you could apply it to neural 
networks. The principle says that the smallest network that can learn the 
training data will be the most successful at predicting the test data. You 
could apply MDL to decision trees too. It says to use the smallest tree that 
explains the training data. You could apply it to any machine learning 
algorithm.

> And, Occam's Razor is not reliable as an axiom of science.  If we were to 
> abide by it we would come to conclusions like a finding that describes an 
> event by saying that "it occurs some of the time,"

If you have a probabilistic model, then you still have to describe the 
difference between its predictions and the actual data. The combined size of 
the model plus the differences is what counts.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Sun, July 4, 2010 12:21:18 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI


I figured out a way to make the Solomonoff Induction iteratively infinite, so I 
guess I was wrong.  Thanks for explaining it to me.  However, I don't accept 
that it is feasible to make those calculations since an examination of the 
infinite programs that could output each individual string would be required.
 
My sense is that the statistics of a examination of a finite number of programs 
that output a finite number of strings could be used in Solomonff Induction to 
to give a reliable probability of what the next bit (or next sequence of bits) 
might be based on the sampling, under the condition that only those cases that 
had previously occurred would occur again and at the same frequencyy during the 
samplings.  However, the attempt to figure the probabilities of concatenation 
of these strings or sub strings would be unreliable and void whatever benefit 
the theoretical model might appear to offer.
 
Logic, probability and compression methods are all useful in AGI even though we 
are constantly violating the laws of logic and probability because it is 
necessary, and we sometimes need to use more complicated models 
(anti-compression so to speak) so that we can consider other possibilities 
based on what we have previously learned.  So, I still don't see how Kolmogrov 
Complexity and Solomonoff Induction are truly useful except as theoretical 
methods that are interesting to consider.
 
And, Occam's Razor is not reliable as an axiom of science.  If we were to abide 
by it we would come to conclusions like a finding that describes an event by 
saying that "it occurs some of the time," since it would be simpler than trying 
to describe the greater circumstances of the event in an effort to see if we 
can find something out about why the event occurred or didn't occur.  In this 
sense Occam's Razor is anti-science since it implies that the status quo should 
be maintained since simpler is better.  All things being equal, simpler is 
better.  I think we all get that.  However, the human mind is capable of re 
weighting the conditions and circumstances of a system to reconsider other 
possibilities and that seems to be an important and necessary method in 
research (and in planning).
 
Jim Bromer  


On Sat, Jul 3, 2010 at 11:39 AM, Matt Mahoney  wrote:

Jim Bromer wrote:
>> You can't assume a priori that the diagonal argument is not relevant. 
>
>
>When I say "infinite" in my proof of Solomonoff induction, I mean countably 
>infinite, as in aleph-null, as in there is a 1 to 1 mapping between the set 
>and N, the set of natural numbers. There are a countably infinite number of 
>finite strings, or of finite programs, or of finite length descriptions of any 
>particular string. For any finite length string or program or description x 
>with nonzero probability, there are a countably infinite number of finite 
>length strings or programs or descriptions that are longer and less likely 
>than x, and a finite number of finite length strings or programs or 
>descriptions that are either shorter or more likely or both than x.
>
>
>Aleph-null is larger than any finite integer. This means that for any finite 
>set and any countably infinite set, there is not a 1 to 1 mapping between the 
>elements, and if you do map all of the elements of the finite set to elements 
>of the infinite set, then there are unmapped elements of the infinite set left 
>over.
>
>

Re: [agi] Reward function vs utility

2010-07-04 Thread Matt Mahoney
Perhaps we now have a better understanding of the risks of uploading to a form 
where we could modify our own software. We already do this to some extent using 
drugs. Evolution will eliminate such failures.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Abram Demski 
To: agi 
Sent: Sun, July 4, 2010 11:43:46 AM
Subject: Re: [agi] Reward function vs utility

Joshua,

But couldn't it game the external utility function by taking actions which 
modify it? For example, if the suggestion is taken literally and you have a 
person deciding the reward at each moment, an AI would want to focus on making 
that person *think* the reward should be high, rather than focusing on actually 
doing well at whatever task it's set...and the two would tend to diverge 
greatly for more and more complex/difficult tasks, since these tend to be 
harder to judge. Furthermore, the AI would be very pleased to knock the human 
out of the loop and push its own buttons. Similar comments would apply to 
automated reward calculations.

--Abram


On Sun, Jul 4, 2010 at 4:40 AM, Joshua Fox  wrote:

Another point. I'm probably repeating the obvious, but perhaps this will be 
useful to some.
>
>
>On the one hand,  an agent could not game a Legg-like intelligence metric by 
>altering the utility function, even an internal one,, since the metric is 
>based on the function before any such change.
>
>
>On the other hand, since an  internally-calculated utility function would 
>necessarily be a function of observations, rather than of actual world state, 
>it could be successfully gamed by altering observations.  
>
>
>This latter objection does not apply to functions which are externally 
>calculated, whether known or unknown.
>
>Joshua
>
>
>
>
>
>
>>
>On Fri, Jul 2, 2010 at 7:23 PM, Joshua Fox  wrote:
>
>>>
>>
>>I found the answer as given by Legg, Machine Superintelligence, p. 72, copied 
>>below. A reward function is used to bypass potential difficulty in 
>>communicating a utility function to the agent.
>>
>>
>>Joshua
>>
>>
>>
>>The existence of a goal raises the problem of how the agent knows what the
>>goal is. One possibility would be for the goal to be known in advance and
>>for this knowledge to be built into the agent. The problem with this is that
>>it limits each agent to just one goal. We need to allow agents that are more
>>flexible, specifically, we need to be able to inform the agent of what the 
>>goal
>>is. For humans this is easily done using language. In general however, the
>>possession of a suffciently high level of language is too strong an assumption
>>to make about the agent. Indeed, even for something as intelligent as a dog
>>or a cat, direct explanation is not very effective.
>>
>>
>>Fortunately there is another possibility which is, in some sense, a blend of
>>the above two. We define an additional communication channel with the sim-
>>plest possible semantics: a signal that indicates how good the agent’s current
>>situation is. We will call this signal the reward. The agent simply has to
>>maximise the amount of reward it receives, which is a function of the goal. In
>>a complex setting the agent might be rewarded for winning a game or solving
>>a puzzle. If the agent is to succeed in its environment, that is, receive a 
>>lot of
>>reward, it must learn about the structure of the environment and in particular
>>what it needs to do in order to get reward.
>>
>>
>>
>>
>>
>>
>>
>>On Mon, Jun 28, 2010 at 1:32 AM, Ben Goertzel  wrote:
>>
>>>>>You can always build the utility function into the assumed universal 
>>>>>Turing machine underlying the definition of algorithmic information...
>>>
>>>I guess this will improve learning rate by some additive constant, in the 
>>>long run ;)
>>>
>>>ben
>>>
>>>
>>>On Sun, Jun 27, 2010 at 4:22 PM, Joshua Fox  wrote:
>>>
>>>This has probably been discussed at length, so I will appreciate a reference 
>>>on this:
>>>>
>>>>
>>>>Why does Legg's definition of intelligence (following on Hutters' AIXI and 
>>>>related work) involve a reward function rather than a utility function? For 
>>>>this purpose, reward is a function of the word state/history which is 
>>>>unknown to the agent while  a utility function is known to the agent. 
>>>>
>>>>
>>>>Even if  we replace the former with the latter, we can still have a 
>>>>definition of intelligence that integra

Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-03 Thread Matt Mahoney
Jim Bromer wrote:
> You can't assume a priori that the diagonal argument is not relevant. 

When I say "infinite" in my proof of Solomonoff induction, I mean countably 
infinite, as in aleph-null, as in there is a 1 to 1 mapping between the set and 
N, the set of natural numbers. There are a countably infinite number of finite 
strings, or of finite programs, or of finite length descriptions of any 
particular string. For any finite length string or program or description x 
with nonzero probability, there are a countably infinite number of finite 
length strings or programs or descriptions that are longer and less likely than 
x, and a finite number of finite length strings or programs or descriptions 
that are either shorter or more likely or both than x.

Aleph-null is larger than any finite integer. This means that for any finite 
set and any countably infinite set, there is not a 1 to 1 mapping between the 
elements, and if you do map all of the elements of the finite set to elements 
of the infinite set, then there are unmapped elements of the infinite set left 
over.

Cantor's diagonalization argument proves that there are infinities larger than 
aleph-null, such as the cardinality of the set of real numbers, which we call 
uncountably infinite. But since I am not using any uncountably infinite sets, I 
don't understand your objection.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Sat, July 3, 2010 9:43:15 AM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI

On Fri, Jul 2, 2010 at 6:08 PM, Matt Mahoney  wrote:

Jim, to address all of your points,
>
>
>Solomonoff induction claims that the probability of a string is proportional 
>to the number of programs that output the string, where each program M is 
>weighted by 2^-|M|. The probability is dominated by the shortest program 
>(Kolmogorov complexity), but it is not exactly the same. The difference is 
>small enough that we may neglect it, just as we neglect differences that 
>depend on choice of language.
 
 
The infinite number of programs that could output the infinite number of 
strings that are to be considered (for example while using Solomonoff induction 
to "predict" what string is being output) lays out the potential for the 
diagonal argument.  You can't assume a priori that the diagonal argument is not 
relevant.  I don't believe that you can prove that it isn't relevant since as 
you say, Kolmogorov Complexity is not computable, and you cannot be sure that 
you have listed all the programs that were able to output a particular string. 
This creates a situation in which the underlying logic of using Solmonoff 
induction is based on incomputable reasoning which can be shown using the 
diagonal argument.
 
This kind of criticism cannot be answered with the kinds of presumptions that 
you used to derive the conclusions that you did.  It has to be answered 
directly.  I can think of other infinity to infinity relations in which the 
potential mappings can be countably derived from the formulas or equations, but 
I have yet to see any analysis which explains why this usage can be.  Although 
you may imagine that the summation of the probabilities can be used just like 
it was an ordinary number, the unchecked usage is faulty.  In other words the 
criticism has to be considered more carefully by someone capable of dealing 
with complex mathematical problems that involve the legitimacy of claims 
between infinite to infinite mappings.
 
Jim Bromer
 
 
 
On Fri, Jul 2, 2010 at 6:08 PM, Matt Mahoney  wrote:

Jim, to address all of your points,
>
>
>Solomonoff induction claims that the probability of a string is proportional 
>to the number of programs that output the string, where each program M is 
>weighted by 2^-|M|. The probability is dominated by the shortest program 
>(Kolmogorov complexity), but it is not exactly the same. The difference is 
>small enough that we may neglect it, just as we neglect differences that 
>depend on choice of language.
>
>
>Here is the proof that Kolmogorov complexity is not computable. Suppose it 
>were. Then I could test the Kolmogorov complexity of strings in increasing 
>order of length (breaking ties lexicographically) and describe "the first 
>string that cannot be described in less than a million bits", contradicting 
>the fact that I just did. (Formally, I could write a program that outputs the 
>first string whose Kolmogorov complexity is at least n bits, choosing n to be 
>larger than my program).
>
>
>Here is the argument that Occam's Razor and Solomonoff distribution must be 
>true. Consider all possible probability distributions p(x) over any infinite 
>set X of possible finite strings x, i.e. any X = {x: p(x) > 0} that is 
>infinite. All such distributions must favor sh

Re: [agi] Re: Huge Progress on the Core of AGI

2010-07-02 Thread Matt Mahoney
Jim, to address all of your points,

Solomonoff induction claims that the probability of a string is proportional to 
the number of programs that output the string, where each program M is weighted 
by 2^-|M|. The probability is dominated by the shortest program (Kolmogorov 
complexity), but it is not exactly the same. The difference is small enough 
that we may neglect it, just as we neglect differences that depend on choice of 
language.

Here is the proof that Kolmogorov complexity is not computable. Suppose it 
were. Then I could test the Kolmogorov complexity of strings in increasing 
order of length (breaking ties lexicographically) and describe "the first 
string that cannot be described in less than a million bits", contradicting the 
fact that I just did. (Formally, I could write a program that outputs the first 
string whose Kolmogorov complexity is at least n bits, choosing n to be larger 
than my program).

Here is the argument that Occam's Razor and Solomonoff distribution must be 
true. Consider all possible probability distributions p(x) over any infinite 
set X of possible finite strings x, i.e. any X = {x: p(x) > 0} that is 
infinite. All such distributions must favor shorter strings over longer ones. 
Consider any x in X. Then p(x) > 0. There can be at most a finite number (less 
than 1/p(x)) of strings that are more likely than x, and therefore an infinite 
number of strings which are less likely than x. Of this infinite set, only a 
finite number (less than 2^|x|) can be shorter than x, and therefore there must 
be an infinite number that are longer than x. So for each x we can partition X 
into 4 subsets as follows:

- shorter and more likely than x: finite
- shorter and less likely than x: finite
- longer and more likely than x: finite
- longer and less likely than x: infinite.

So in this sense, any distribution over the set of strings must favor shorter 
strings over longer ones.

-- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Fri, July 2, 2010 4:09:38 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI




On Fri, Jul 2, 2010 at 2:25 PM, Jim Bromer  wrote:  
There cannot be a one to one correspondence to the representation of the 
shortest program to produce a string and the strings that they produce.  This 
means that if the consideration of the hypotheses were to be put into general 
mathematical form it must include the potential of many to one relations 
between candidate programs (or subprograms) and output strings.
 
But, there is also no way to determine what the "shortest" program is, since 
there may be different programs that are the same length.  That means that 
there is a many to one relation between programs and program "length".  So the 
claim that you could just iterate through programs by length is false.  This is 
the goal of algorithmic information theory not a premise of a methodology that 
can be used.  So you have the diagonalization problem. 
 
A counter argument is that there are only a finite number of Turing Machine 
programs of a given length.  However, since you guys have specifically 
designated that this theorem applies to any construction of a Turing Machine it 
is not clear that this counter argument can be used.  And there is still the 
specific problem that you might want to try a program that writes a longer 
program to output a string (or many strings).  Or you might want to write a 
program that can be called to write longer programs on a dynamic basis.  I 
think these cases, where you might consider a program that outputs a longer 
program, (or another instruction string for another Turing Machine) constitutes 
a serious problem, that at the least, deserves to be answered with sound 
analysis.
 
Part of my original intuitive argument, that I formed some years ago, was that 
without a heavy constraint on the instructions for the program, it will be 
practically impossible to test or declare that some program is indeed the 
shortest program.  However, I can't quite get to the point now that I can say 
that there is definitely a diagonalization problem.
 
Jim Bromer

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] masterpiece on an iPad

2010-07-02 Thread Matt Mahoney
An AGI only has to predict your behavior so that it can serve you better by 
giving you what you want without you asking for it. It is not a copy of your 
mind. It is a program that can call a function that simulates your mind for 
some arbitrary purpose determined by its programmer.

 -- Matt Mahoney, matmaho...@yahoo.com





From: John G. Rose 
To: agi 
Sent: Fri, July 2, 2010 11:39:23 AM
Subject: RE: [agi] masterpiece on an iPad


An AGI may not really think like we do, it may just execute code. 
 
Though I suppose you could program a lot of fuzzy loops and idle speculation, 
entertaining possibilities, having human "think" envy.. 
 
John
 
From:Matt Mahoney [mailto:matmaho...@yahoo.com] 
Sent: Friday, July 02, 2010 8:21 AM
To: agi
Subject: Re: [agi] masterpiece on an iPad
 
AGI is all about building machines that think, so you don't have to.

 
-- Matt Mahoney, matmaho...@yahoo.com
 
 



From:Mike Tintner 
To: agi 
Sent: Fri, July 2, 2010 9:37:51 AM
Subject: Re: [agi] masterpiece on an iPad
that's like saying cartography or cartoons could be done a lot faster if they 
just used cameras -  ask Michael to explain what the hand can draw that the 
camera can't
 
From:Matt Mahoney 
Sent:Friday, July 02, 2010 2:21 PM
To:agi 
Subject:Re: [agi] masterpiece on an iPad
 
It could be done a lot faster if the iPad had a camera.

 
-- Matt Mahoney, matmaho...@yahoo.com 
 
 



From:Mike Tintner 
To: agi 
Sent: Fri, July 2, 2010 6:28:58 AM
Subject: [agi] masterpiece on an iPad
http://www.telegraph.co.uk/culture/culturevideo/artvideo/7865736/Artist-creates-masterpiece-on-an-iPad.html
 
McLuhan argues that touch is the central sense - the one that binds the others. 
He may be right. The i-devices integrate touch into intelligence.
agi| Archives | Modify Your Subscription  
agi| Archives| Modify Your Subscription  
 
agi| Archives| Modify Your Subscription  
agi| Archives| ModifyYour Subscription  
 
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] masterpiece on an iPad

2010-07-02 Thread Matt Mahoney
AGI is all about building machines that think, so you don't have to.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Fri, July 2, 2010 9:37:51 AM
Subject: Re: [agi] masterpiece on an iPad


that's like saying cartography or 
cartoons could be done a lot faster if they just used cameras -  ask 
Michael to explain what the hand can draw that the 
camera can't


From: Matt Mahoney 
Sent: Friday, July 02, 2010 2:21 PM
To: agi 
Subject: Re: [agi] masterpiece on an iPad

It could be done a lot faster if the iPad had a camera.

 -- Matt Mahoney, matmaho...@yahoo.com 





 From: Mike Tintner 
To: agi 
Sent: Fri, July 2, 2010 6:28:58 
AM
Subject: [agi] masterpiece 
on an iPad


http://www.telegraph.co.uk/culture/culturevideo/artvideo/7865736/Artist-creates-masterpiece-on-an-iPad.html
 
McLuhan argues that touch is the central sense - 
the one that binds the others. He may be right. The i-devices integrate touch 
into intelligence.
agi | Archives  | Modify Your Subscription  
agi | Archives  | Modify Your Subscription   
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] masterpiece on an iPad

2010-07-02 Thread Matt Mahoney
It could be done a lot faster if the iPad had a camera.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Fri, July 2, 2010 6:28:58 AM
Subject: [agi] masterpiece on an iPad


http://www.telegraph.co.uk/culture/culturevideo/artvideo/7865736/Artist-creates-masterpiece-on-an-iPad.html
 
McLuhan argues that touch is the central sense - 
the one that binds the others. He may be right. The i-devices integrate touch 
into intelligence.
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Huge Progress on the Core of AGI

2010-06-30 Thread Matt Mahoney
Jim, what evidence do you have that Occam's Razor or algorithmic information 
theory is wrong, besides your own opinions? It is well established that elegant 
(short) theories are preferred in all branches of science because they have 
greater predictive power.

Also, what does this have to do with Cantor's diagonalization argument? AIT 
considers only the countably infinite set of hypotheses.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Jim Bromer 
To: agi 
Sent: Wed, June 30, 2010 9:13:44 AM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI


On Tue, Jun 29, 2010 at 11:46 PM, Abram Demski  wrote:
In brief, the answer to your question is: we formalize the description length 
heuristic by assigning lower probabilities to longer hypotheses, and we apply 
Bayes law to update these probabilities given the data we observe. This 
updating captures the idea that we should reward theories which explain/expect 
more of the observations; it also provides a natural way to balance simplicity 
vs explanatory power, so that we can compare any two theories with a single 
scoring mechanism. Bayes Law automatically places the right amount of pressure 
to avoid overly elegant explanations which don't get much right, and to avoid 
overly complex explanations which fit the observations perfectly but which 
probably won't generalize to new data.
...
If you go down this path, you will eventually come to understand (and, 
probably, accept) algorithmic information theory. Matt may be tring to force it 
on you too soon. :)
--Abram 
 
David was asking about theories of explanation, and here you are suggesting 
that following a certain path of reasoning will lead to accepting AIT.  What 
nonsense.  Even assuming that Baye's law can be used to update probabilities of 
idealized utility, the connection between description length and explanatory 
power in general AI is tenuous.  And when you realize that AIT is an 
unattainable idealism that lacks mathematical power (I do not believe that it 
is a valid mathematical method because it is incomputable and therefore 
innumerable and cannot be used to derive probability distributions even as 
ideals) you have to accept that the connection between explanatory theories and 
AIT is not established except as a special case based on the imagination that a 
similarities between a subclass of practical examples is the same as a powerful 
generalization of those examples.  
 
The problem is that while compression seems to be related to intelligence, it 
is not equivalent to intelligence.  A much stronger but similarly false 
argument is that memory is intelligence.  Of course memory is a major part of 
intelligence, but it is not everything.  The argument that AIT is a reasonable 
substitute for developing more sophisticated theories about conceptual 
explanation is not well founded, it lacks any experimental evidence other than 
a spattering of results on simplistic cases, and it is just wrong to suggest 
that there is no reason to consider other theories of explanation.
 
Yes compression has something to do with intelligence and, in some special 
cases it can be shown to act as an idealism for numerical rationality.  And yes 
unattainable theories that examine the boundaries of productive mathematical 
systems is a legitimate subject for mathematics.  But there is so much more to 
theories of explanatory reasoning that I genuinely feel sorry for those of you, 
who originally motivated to develop better AGI programs, would get caught in 
the obvious traps of AIT and AIXI.
 
Jim Bromer 

 
On Tue, Jun 29, 2010 at 11:46 PM, Abram Demski  wrote:

David,
>
>What Matt is trying to explain is all right, but I think a better way of 
>answering your question would be to invoke the mighty mysterious Bayes' Law.
>
>I had an epiphany similar to yours (the one that started this thread) about 5 
>years ago now. At the time I did not know that it had all been done before. I 
>think many people feel this way about MDL. Looking into the MDL (minimum 
>description length) literature would be a good starting point.
>
>In brief, the answer to your question is: we formalize the description length 
>heuristic by assigning lower probabilities to longer hypotheses, and we apply 
>Bayes law to update these probabilities given the data we observe. This 
>updating captures the idea that we should reward theories which explain/expect 
>more of the observations; it also provides a natural way to balance simplicity 
>vs explanatory power, so that we can compare any two theories with a single 
>scoring mechanism. Bayes Law automatically places the right amount of pressure 
>to avoid overly elegant explanations which don't get much right, and to avoid 
>overly complex explanations which fit the observations perfectly but which 
>probably won't generalize to new data.
>
>Bayes

Re: [agi] Re: Huge Progress on the Core of AGI

2010-06-29 Thread Matt Mahoney
You can always find languages that favor either hypothesis. Suppose that you 
want to predict the sequence 10, 21, 32, ? and we write our hypothesis as a 
function that takes the trial number (0, 1, 2, 3...) and returns the outcome. 
The sequence 10, 21, 32, 43, 54... would be coded:

int hypothesis_1(int trial) {
  return trial*11+10;
}

The sequence 10, 21, 32, 10, 21, 32... would be coded

int hypothesis_2(int trial) {
  return trial%3*11+10;
}

which is longer and therefore less likely.

Here is another example: predict the sequence 0, 1, 4, 9, 16, 25, 36, 49, ?

Can you find a program shorter than this that doesn't predict 64?

int hypothesis_1(int trial) {
  return trial*trial;
}

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 3:48:01 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI

Such an example is no where near sufficient to accept the assertion that 
program size is the right way to define simplicity of a hypothesis.

Here is a counter example. It requires a slightly more complex example because 
all zeros doesn't leave any room for alternative hypotheses.

Here is the sequence: 10, 21, 32

void hypothesis_1() {
   int ten = 10;
   int counter = 0;

while (1)
{
   print(ten+counter)
   ten = ten + 10;
   counter = counter + 1;
}

  }

void hypothesis_2() {

while (1)
   print("10 21 32")

  }

Hypothesis 2 is simpler, yet clearly wrong. These examples don't really show 
anything.

Dave


On Tue, Jun 29, 2010 at 3:15 PM, Matt Mahoney  wrote:

David Jones wrote:
>> I really don't think this is the right way to calculate simplicity. 
>
>
>I will give you an example, because examples are more convincing than proofs.
>
>
>Suppose you perform a sequence of experiments whose outcome can either be 0 or 
>1. In the first 10 trials you observe 00. What do you expect to 
>observe in the next trial?
>
>
>Hypothesis 1: the outcome is always 0.
>Hypothesis 2: the outcome is 0 for the first 10 trials and 1 thereafter.
>
>
>Hypothesis 1 is shorter than 2, so it is more likely to be correct.
>
>
>If I describe
> the two hypotheses in French or Chinese, then 1 is still shorter than 2.
>
>
>If I describe the two hypotheses in C, then 1 is shorter than 2.
>
>
>  void hypothesis_1() {
>>while (1) printf("0");
>  }
>
>
>  void hypothesis_2() {
>int i;
>for (i=0; i<10; ++i) printf("0");
>while (1) printf("1");
>  }
>
>
>If I translate these programs into Perl or Lisp or x86 assembler, then 1 will 
>still be shorter than 2.
>
>
>I realize there might be smaller equivalent programs. But I think you could 
>find a smaller program equivalent to hypothesis_1 than hypothesis_2.
>
>
>I realize there are other hypotheses than 1 or 2. But I think that the 
>smallest one you can find that outputs
> eleven bits of which the first ten are zeros will be a program that outputs 
> another zero.
>
>
>I realize that you could rewrite 1 so that it is longer than 2. But it is the 
>shortest version that counts. More specifically consider all programs in which 
>the first 10 outputs are 0. Then weight each program by 2^-length. So the 
>shortest programs dominate.
>
>
>I realize you could make up a language where the shortest encoding of 
>hypothesis 2 is shorter than 1. You could do this for any pair of hypotheses. 
>However, I think if you stick to "simple" languages (and I realize this is a 
>circular definition), then 1 will usually be shorter than 2.
>
> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: David Jones 
>To: agi 
>Sent: Tue, June 29, 2010 1:31:01 PM
>
>Subject: Re: [agi] Re: Huge Progress on the Core of AGI
>
>
>
>
>
>>On Tue, Jun 29, 2010 at 11:26 AM, Matt Mahoney  wrote:
>
>>>
>>> Right. But Occam's Razor is not complete. It says simpler is better, but 1) 
>>> this only applies when two hypotheses have the same explanatory power and 
>>> 2) what defines simpler? 
>>
>>
>>A hypothesis is a program that outputs the observed data. It "explains" the 
>>data if its output matches what is observed. The "simpler" hypothesis is the 
>>shorter program, measured in bits.
>
>I can't be confident that bits is the right way to do it. I suspect bits is an 
>approximation of a more accurate method. I also suspect that you can write a 
>more complex explanation "program" with the same number of bits. So, there are 
>some flaws with this approach. It is an interesting idea to consider though. 
>>
> 
>
>

Re: [agi] A Primary Distinction for an AGI

2010-06-29 Thread Matt Mahoney
Answering questions is the same problem as predicting the answers. If you can 
compute p(A|Q) where Q is the question (and previous context of the 
conversation) and A is the answer, then you can also choose an answer A from 
the same distribution. If p() correctly models human communication, then the 
response would be indistinguishable from a human in a Turing test.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 3:43:53 PM
Subject: Re: [agi] A Primary Distinction for an AGI

the purpose of text is to convey something. It has to be interpreted. who cares 
about predicting the next word if you can't interpret a single bit of it.


On Tue, Jun 29, 2010 at 3:43 PM, David Jones  wrote:

People do not predict the next words of text. We anticipate it, but when 
something different shows up, we accept it if it is *explanatory*. Using 
compression like algorithms though will never be able to do this type of 
explanatory reasoning, which is required to disambiguate text. It is certainly 
not sufficient for learning language, which is not at all about predicting text.
>
>
>
>On Tue, Jun 29, 2010 at 3:38 PM, Matt Mahoney  wrote:
>
>>>
>>Experiments in text compression show that text alone is sufficient for 
>>learning to predict text.
>>
>>
>>I realize that for a machine to pass the Turing test, it needs a visual model 
>>of the world. Otherwise it would have a hard time with questions like "what 
>>word in this ernai1 did I spell wrong"? Obviously the easiest way to build a 
>>visual model is with vision, but it is not the only way.
>>
>> -- Matt Mahoney, matmaho...@yahoo.com
>>
>>
>>
>>
>>

From: David Jones
>> 
>>To: agi 
>>Sent: Tue, June 29, 2010 3:22:33 PM
>>
>>Subject: Re: [agi] A Primary Distinction for an AGI
>>
>>
>>I certainly agree that the techniques and explanation generating algorithms 
>>for learning language are hard coded into our brain. But, those techniques 
>>alone are not sufficient to learn language in the absence of sensory 
>>perception or some other way of getting the data required.
>>
>>Dave
>>
>>
>>On Tue, Jun 29, 2010 at 3:19 PM, Matt Mahoney  wrote:
>>
>>David Jones wrote:
>>>>  The knowledge for interpreting language though should not be 
>>>> pre-programmed. 
>>>
>>>
>>>I think that human brains are wired differently than other animals to make 
>>>language learning easier. We have not been successful in training other 
>>>primates to speak, even though they have all the right anatomy such as vocal 
>>>chords, tongue, lips, etc. When primates have been taught sign language, 
>>>they have not successfully mastered forming sentences.
>>>
>>> -- Matt Mahoney, matmaho...@yahoo.com
>>>
>>>
>>>>>>
>>>
>>>
>>>

From: David Jones 
>>>To: agi 
>>>Sent: Tue, June 29, 2010 3:00:09 PM
>>>>>>
>>>
>>>
>>>Subject: Re: [agi] A Primary Distinction for an AGI
>>>
>>>
>>>The point I was trying to make is that an approach that tries to interpret 
>>>language just using language itself and without sufficient information or 
>>>the means to realistically acquire that information, *should* fail. 
>>>
>>>>>>On the other hand, an approach that tries to interpret vision with 
>>>>>>minimal upfront knowledge needs *should* succeed because the knowledge 
>>>>>>required to automatically learn to interpret images is amenable to 
>>>>>>preprogramming. In addition, such knowledge must be pre-programmed. The 
>>>>>>knowledge for interpreting language though should not be pre-programmed. 
>>>
>>>Dave
>>>
>>>
>>>On Tue, Jun 29, 2010 at 2:51 PM, Matt Mahoney  wrote:
>>>
>>>David Jones wrote:
>>>>> I wish people understood this better.
>>>>
>>>>
>>>>For example, animals can be intelligent even though they lack language 
>>>>because they can see. True, but an AGI with language skills is more useful 
>>>>than one without.
>>>>
>>>>
>>>>And yes, I realize that language, vision, motor skills, hearing, and all 
>>>>the other senses and outputs are tied together. Skills in any area make 
>>>>learning the others easier.
>>>>
>>>>>>>> -- Matt Maho

Re: [agi] A Primary Distinction for an AGI

2010-06-29 Thread Matt Mahoney
Experiments in text compression show that text alone is sufficient for learning 
to predict text.

I realize that for a machine to pass the Turing test, it needs a visual model 
of the world. Otherwise it would have a hard time with questions like "what 
word in this ernai1 did I spell wrong"? Obviously the easiest way to build a 
visual model is with vision, but it is not the only way.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 3:22:33 PM
Subject: Re: [agi] A Primary Distinction for an AGI

I certainly agree that the techniques and explanation generating algorithms for 
learning language are hard coded into our brain. But, those techniques alone 
are not sufficient to learn language in the absence of sensory perception or 
some other way of getting the data required.

Dave


On Tue, Jun 29, 2010 at 3:19 PM, Matt Mahoney  wrote:

David Jones wrote:
>>  The knowledge for interpreting language though should not be 
>> pre-programmed. 
>
>
>I think that human brains are wired differently than other animals to make 
>language learning easier. We have not been successful in training other 
>primates to speak, even though they have all the right anatomy such as vocal 
>chords, tongue, lips, etc. When primates have been taught sign language, they 
>have not successfully mastered forming sentences.
>
> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: David Jones 
>To: agi 
>Sent: Tue, June 29, 2010 3:00:09 PM
>
>Subject: Re: [agi] A Primary Distinction for an AGI
>
>
>The point I was trying to make is that an approach that tries to interpret 
>language just using language itself and without sufficient information or the 
>means to realistically acquire that information, *should* fail. 
>
>>On the other hand, an approach that tries to interpret vision with minimal 
>>upfront knowledge needs *should* succeed because the knowledge required to 
>>automatically learn to interpret images is amenable to preprogramming. In 
>>addition, such knowledge must be pre-programmed. The knowledge for 
>>interpreting language though should not be pre-programmed. 
>
>Dave
>
>
>On Tue, Jun 29, 2010 at 2:51 PM, Matt Mahoney  wrote:
>
>David Jones wrote:
>>> I wish people understood this better.
>>
>>
>>For example, animals can be intelligent even though they lack language 
>>because they can see. True, but an AGI with language skills is more useful 
>>than one without.
>>
>>
>>And yes, I realize that language, vision, motor skills, hearing, and all the 
>>other senses and outputs are tied together. Skills in any area make learning 
>>the others easier.
>>
>>>> -- Matt Mahoney, matmaho...@yahoo.com
>>
>>
>>
>>
>>

From: David Jones 
>>To: agi 
>>Sent: Tue, June 29, 2010 1:42:51 PM
>>>>
>>
>>Subject: Re: [agi] A Primary Distinction for an AGI
>>
>>
>>Mike, 
>>
>>THIS is the flawed reasoning that causes people to ignore vision as the right 
>>way to create AGI. And I've finally come up with a great way to show you how 
>>wrong this reasoning is. 
>>
>>I'll give you an extremely obvious argument that proves that vision requires 
>>much less knowledge to interpret than language does. Let's say that you have 
>>never been to egypt, you have never seen some particular movie before.  But
>> if you see the movie, an alien landscape, an alien world, a new place or any 
>> such new visual experience, you can immediately interpret it in terms of 
>> spacial, temporal, compositional and other relationships. 
>>
>>Now, go to egypt and listen to them speak. Can you interpret it? Nope. Why?! 
>>Because you don't have enough information. The language itself does not 
>>contain any information to help you interpret it. We do not learn language 
>>simply by listening. We learn based on evidence from how the language is used 
>>and how it occurs in our daily lives. Without that experience, you cannot 
>>interpret it.
>>
>>But with vision, you do not need extra knowledge to interpret a new 
>>situation. You can recognize completely new objects without any training 
>>except for simply observing them in their natural state. 
>>
>>I wish people understood this better.
>>
>>Dave
>>
>>
>>On Tue, Jun 29, 2010 at 12:51 PM, Mike Tintner  
>>wrote:
>>
>>>>>
>>>
>>>
>>>
>>>Just off the cuff here - isn't the same true for 
>>>visi

Re: [agi] A Primary Distinction for an AGI

2010-06-29 Thread Matt Mahoney
David Jones wrote:
>  The knowledge for interpreting language though should not be pre-programmed. 

I think that human brains are wired differently than other animals to make 
language learning easier. We have not been successful in training other 
primates to speak, even though they have all the right anatomy such as vocal 
chords, tongue, lips, etc. When primates have been taught sign language, they 
have not successfully mastered forming sentences.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 3:00:09 PM
Subject: Re: [agi] A Primary Distinction for an AGI

The point I was trying to make is that an approach that tries to interpret 
language just using language itself and without sufficient information or the 
means to realistically acquire that information, *should* fail. 

On the other hand, an approach that tries to interpret vision with minimal 
upfront knowledge needs *should* succeed because the knowledge required to 
automatically learn to interpret images is amenable to preprogramming. In 
addition, such knowledge must be pre-programmed. The knowledge for interpreting 
language though should not be pre-programmed. 

Dave


On Tue, Jun 29, 2010 at 2:51 PM, Matt Mahoney  wrote:

David Jones wrote:
>> I wish people understood this better.
>
>
>For example, animals can be intelligent even though they lack language because 
>they can see. True, but an AGI with language skills is more useful than one 
>without.
>
>
>And yes, I realize that language, vision, motor skills, hearing, and all the 
>other senses and outputs are tied together. Skills in any area make learning 
>the others easier.
>
>> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: David Jones 
>To: agi 
>Sent: Tue, June 29, 2010 1:42:51 PM
>
>Subject: Re: [agi] A Primary Distinction for an AGI
>
>
>Mike, 
>
>THIS is the flawed reasoning that causes people to ignore vision as the right 
>way to create AGI. And I've finally come up with a great way to show you how 
>wrong this reasoning is. 
>
>I'll give you an extremely obvious argument that proves that vision requires 
>much less knowledge to interpret than language does. Let's say that you have 
>never been to egypt, you have never seen some particular movie before.  But
> if you see the movie, an alien landscape, an alien world, a new place or any 
> such new visual experience, you can immediately interpret it in terms of 
> spacial, temporal, compositional and other relationships. 
>
>Now, go to egypt and listen to them speak. Can you interpret it? Nope. Why?! 
>Because you don't have enough information. The language itself does not 
>contain any information to help you interpret it. We do not learn language 
>simply by listening. We learn based on evidence from how the language is used 
>and how it occurs in our daily lives. Without that experience, you cannot 
>interpret it.
>
>But with vision, you do not need extra knowledge to interpret a new situation. 
>You can recognize completely new objects without any training except for 
>simply observing them in their natural state. 
>
>I wish people understood this better.
>
>Dave
>
>
>On Tue, Jun 29, 2010 at 12:51 PM, Mike Tintner  
>wrote:
>
>>>
>>
>>
>>
>>Just off the cuff here - isn't the same true for 
>>vision? You can't learn vision from vision. Just as all NLP has no connection 
>>with the real world, and totally relies on the human programmer's knowledge 
>>of 
>>that world. 
>> 
>>Your visual program actually relies totally on your 
>>visual "vocabulary" - not its own. That is the inevitable penalty of 
>>processing 
>>unreal signals on a computer screen which are not in fact connected to the 
>>real world any more than the verbal/letter signals involved in NLP 
>>are.
>> 
>>What you need to do - what anyone in your situation 
>>with anything like your asprations needs to do - is to hook up with a 
>>roboticist. Everyone here should be doing that.
>> 
>>
>>
>>From: David Jones 
>>Sent: Tuesday, June 29, 2010 5:27 PM
>>To: agi 
>>Subject: Re: [agi] A Primary Distinction for an 
>>AGI
>>
>>
>>You can't learn language from language without embedding way more knowledge 
>>than is reasonable. Language does not contain the information required for 
>>its 
>>interpretation. There is no *reason* to interpret the language into any of 
>>the 
>>infinite possible interpretaions. There is nothing to explain but it requires 
>>explanatory reasoning to determine the correct real 

Re: [agi] Re: Huge Progress on the Core of AGI

2010-06-29 Thread Matt Mahoney
David Jones wrote:
> I really don't think this is the right way to calculate simplicity. 

I will give you an example, because examples are more convincing than proofs.

Suppose you perform a sequence of experiments whose outcome can either be 0 or 
1. In the first 10 trials you observe 00. What do you expect to observe 
in the next trial?

Hypothesis 1: the outcome is always 0.
Hypothesis 2: the outcome is 0 for the first 10 trials and 1 thereafter.

Hypothesis 1 is shorter than 2, so it is more likely to be correct.

If I describe the two hypotheses in French or Chinese, then 1 is still shorter 
than 2.

If I describe the two hypotheses in C, then 1 is shorter than 2.

  void hypothesis_1() {
while (1) printf("0");
  }

  void hypothesis_2() {
int i;
for (i=0; i<10; ++i) printf("0");
while (1) printf("1");
  }

If I translate these programs into Perl or Lisp or x86 assembler, then 1 will 
still be shorter than 2.

I realize there might be smaller equivalent programs. But I think you could 
find a smaller program equivalent to hypothesis_1 than hypothesis_2.

I realize there are other hypotheses than 1 or 2. But I think that the smallest 
one you can find that outputs eleven bits of which the first ten are zeros will 
be a program that outputs another zero.

I realize that you could rewrite 1 so that it is longer than 2. But it is the 
shortest version that counts. More specifically consider all programs in which 
the first 10 outputs are 0. Then weight each program by 2^-length. So the 
shortest programs dominate.

I realize you could make up a language where the shortest encoding of 
hypothesis 2 is shorter than 1. You could do this for any pair of hypotheses. 
However, I think if you stick to "simple" languages (and I realize this is a 
circular definition), then 1 will usually be shorter than 2.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 1:31:01 PM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI




On Tue, Jun 29, 2010 at 11:26 AM, Matt Mahoney  wrote:

> Right. But Occam's Razor is not complete. It says simpler is better, but 1) 
> this only applies when two hypotheses have the same explanatory power and 2) 
> what defines simpler? 
>
>
>A hypothesis is a program that outputs the observed data. It "explains" the 
>data if its output matches what is observed. The "simpler" hypothesis is the 
>shorter program, measured in bits.

I can't be confident that bits is the right way to do it. I suspect bits is an 
approximation of a more accurate method. I also suspect that you can write a 
more complex explanation "program" with the same number of bits. So, there are 
some flaws with this approach. It is an interesting idea to consider though. 
 


>
>The language used to describe the data can be any Turing complete programming 
>language (C, Lisp, etc) or any natural language such as English. It does not 
>matter much which language you use, because for any two languages there is a 
>fixed length procedure, described in either of the languages, independent of 
>the data, that translates descriptions in
> one language to the other.

Hypotheses don't have to be written in actual computer code and probably 
shouldn't be because hypotheses are not really meant to be "run" per say. And 
outputs are not necessarily the right way to put it either. Outputs imply 
prediction. And as mike has often pointed out, things cannot be precisely 
predicted. We can, however, determine whether a particular observation fits 
expectations, rather than equals some prediction. There may be multiple 
possible outcomes that we expect and which would be consistent with a 
hypothesis, which is why actual prediction should not be used.


> For example, the simplest hypothesis for all visual interpretation is that 
> everything in the first image is gone in the second image, and everything in 
> the second image is a new object. Simple. Done. Solved :) right? 
>
>
>The hypothesis is not the simplest. The program that outputs the two frames as 
>if independent cannot be smaller than the two frames compressed independently. 
>The program could be made smaller if it only described how the second frame is 
>different than the first. It would be more likely to correctly predict the 
>third frame if it continued to run and described how it would be different 
>than the second frame.

I really don't think this is the right way to calculate simplicity. 
 


>
>> I don't think much progress has been made in this area, but I'd like to know 
>> what other people have done and any successes they've had.
>
>
>Kolmogorov proved that the solution is not
> computable. Given a hypothesis (a description 

Re: [agi] A Primary Distinction for an AGI

2010-06-29 Thread Matt Mahoney
David Jones wrote:
> I wish people understood this better.

For example, animals can be intelligent even though they lack language because 
they can see. True, but an AGI with language skills is more useful than one 
without.

And yes, I realize that language, vision, motor skills, hearing, and all the 
other senses and outputs are tied together. Skills in any area make learning 
the others easier.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 1:42:51 PM
Subject: Re: [agi] A Primary Distinction for an AGI

Mike, 

THIS is the flawed reasoning that causes people to ignore vision as the right 
way to create AGI. And I've finally come up with a great way to show you how 
wrong this reasoning is. 

I'll give you an extremely obvious argument that proves that vision requires 
much less knowledge to interpret than language does. Let's say that you have 
never been to egypt, you have never seen some particular movie before.  But if 
you see the movie, an alien landscape, an alien world, a new place or any such 
new visual experience, you can immediately interpret it in terms of spacial, 
temporal, compositional and other relationships. 

Now, go to egypt and listen to them speak. Can you interpret it? Nope. Why?! 
Because you don't have enough information. The language itself does not contain 
any information to help you interpret it. We do not learn language simply by 
listening. We learn based on evidence from how the language is used and how it 
occurs in our daily lives. Without that experience, you cannot interpret it.

But with vision, you do not need extra knowledge to interpret a new situation. 
You can recognize completely new objects without any training except for simply 
observing them in their natural state. 

I wish people understood this better.

Dave


On Tue, Jun 29, 2010 at 12:51 PM, Mike Tintner  wrote:

>
>
>
>
>Just off the cuff here - isn't the same true for 
>vision? You can't learn vision from vision. Just as all NLP has no connection 
>with the real world, and totally relies on the human programmer's knowledge of 
>that world. 
> 
>Your visual program actually relies totally on your 
>visual "vocabulary" - not its own. That is the inevitable penalty of 
>processing 
>unreal signals on a computer screen which are not in fact connected to the 
>real world any more than the verbal/letter signals involved in NLP 
>are.
> 
>What you need to do - what anyone in your situation 
>with anything like your asprations needs to do - is to hook up with a 
>roboticist. Everyone here should be doing that.
> 
>
>
>From: David Jones 
>Sent: Tuesday, June 29, 2010 5:27 PM
>To: agi 
>Subject: Re: [agi] A Primary Distinction for an 
>AGI
>
>
>You can't learn language from language without embedding way more knowledge 
>than is reasonable. Language does not contain the information required for its 
>interpretation. There is no *reason* to interpret the language into any of the 
>infinite possible interpretaions. There is nothing to explain but it requires 
>explanatory reasoning to determine the correct real world interpretation
>On Jun 29, 2010 10:58 AM, "Matt Mahoney"  wrote:
>>
>>
>>David Jones wrote:
>>> Natural language 
>>  requires more than the words on the page in the real world. Of...
>>Any knowledge that can be demonstrated over a 
>>  text-only channel (as in the Turing test) can also be learned over a 
>> text-only 
>>  channel.
>>
>>
>>> Cyc also is trying to store knowledge 
>>  about a super complicated world in simplistic forms and al...
>>Cyc failed because it lacks natural language. The vast knowledge 
>>  store of the internet is unintelligible to Cyc. The average person can't 
>>  use it because they don't speak Cycl and because they have neither the 
>> ability 
>>  nor the patience to translate their implicit thoughts into augmented first 
>>  order logic. Cyc's approach was understandable when they started in 1984 
>> when 
>>  they had neither the internet nor the vast computing power that is required 
>> to 
>>  learn natural language from unlabeled examples like children do.
>>
>>
>>> Vision and other sensory interpretaion, on 
>>  the other hand, do not require more info because that...
>>Without natural language, your system will fail too. You don't have 
>>  enough computing power to learn language, much less the million times more 
>>  computing power you need to learn to see.
>>
>>
>> 
>>-- Matt Mahoney, matmaho...@yahoo.com
>>
>>
>>
>>_

Re: [agi] Re: Huge Progress on the Core of AGI

2010-06-29 Thread Matt Mahoney
> Right. But Occam's Razor is not complete. It says simpler is better, but 1) 
> this only applies when two hypotheses have the same explanatory power and 2) 
> what defines simpler? 

A hypothesis is a program that outputs the observed data. It "explains" the 
data if its output matches what is observed. The "simpler" hypothesis is the 
shorter program, measured in bits.

The language used to describe the data can be any Turing complete programming 
language (C, Lisp, etc) or any natural language such as English. It does not 
matter much which language you use, because for any two languages there is a 
fixed length procedure, described in either of the languages, independent of 
the data, that translates descriptions in one language to the other.

> For example, the simplest hypothesis for all visual interpretation is that 
> everything in the first image is gone in the second image, and everything in 
> the second image is a new object. Simple. Done. Solved :) right? 

The hypothesis is not the simplest. The program that outputs the two frames as 
if independent cannot be smaller than the two frames compressed independently. 
The program could be made smaller if it only described how the second frame is 
different than the first. It would be more likely to correctly predict the 
third frame if it continued to run and described how it would be different than 
the second frame.

> I don't think much progress has been made in this area, but I'd like to know 
> what other people have done and any successes they've had.

Kolmogorov proved that the solution is not computable. Given a hypothesis (a 
description of the observed data, or a program that outputs the observed data), 
there is no general procedure or test to determine whether a shorter (simpler, 
better) hypothesis exists. Proof: suppose there were. Then I could describe 
"the first data set that cannot be described in less than a million bits" even 
though I just did. (By "first" I mean the first data set encoded by a string 
from shortest to longest, breaking ties lexicographically).

That said, I believe the state of the art in both language and vision are based 
on hierarchical neural models, i.e. pattern recognition using learned weighted 
combinations of simpler patterns. I am more familiar with language. The top 
ranked programs can be found at http://mattmahoney.net/dc/text.html

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 10:44:41 AM
Subject: Re: [agi] Re: Huge Progress on the Core of AGI

Thanks Matt,

Right. But Occam's Razor is not complete. It says simpler is better, but 1) 
this only applies when two hypotheses have the same explanatory power and 2) 
what defines simpler? 

So, maybe what I want to know from the state of the art in research is: 

1) how precisely do other people define "simpler"
and
2) More importantly, how do you compare competing explanations/hypotheses that 
have more or less explanatory power. Simpler does not apply unless you are 
comparing equally explanatory hypotheses. 

For example, the simplest hypothesis for all visual interpretation is that 
everything in the first image is gone in the second image, and everything in 
the second image is a new object. Simple. Done. Solved :) right? Well, clearly 
a more complicated explanation is warranted because a more complicated 
explanation is more *explanatory* and a better explanation. So, why is it 
better? Can it be defined as better in a precise way so that you can compare 
arbitrary hypotheses or explanations? That is what I'm trying to learn about. I 
don't think much progress has been made in this area, but I'd like to know what 
other people have done and any successes they've had.

Dave



On Tue, Jun 29, 2010 at 10:29 AM, Matt Mahoney  wrote:

David Jones wrote:
>> If anyone has any knowledge of or references to the state of the art in 
>> explanation-based reasoning, can you send me keywords or links? 
>
>
>The simplest explanation of the past is the best predictor of the future.
>http://en.wikipedia.org/wiki/Occam's_razor
>http://www.scholarpedia.org/article/Algorithmic_probability
>
> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: David Jones 
>
>To: agi 
>Sent: Tue, June 29, 2010 9:05:45 AM
>Subject: [agi] Re: Huge Progress on the Core of AGI
>
>
>If anyone has any knowledge of or references to the state of the art in 
>explanation-based reasoning, can you send me keywords or links? I've read some 
>through google, but I'm not really satisfied with anything I've found. 
>
>Thanks,
>
>Dave
>
>
>On Sun, Jun 27, 2010 at 1:31 AM, David Jones  wrote:
>
>>>A method for comparing hypotheses in ex

Re: [agi] A Primary Distinction for an AGI

2010-06-29 Thread Matt Mahoney
David Jones wrote:
> Natural language requires more than the words on the page in the real world. 
> Of course that didn't work.

Any knowledge that can be demonstrated over a text-only channel (as in the 
Turing test) can also be learned over a text-only channel.

> Cyc also is trying to store knowledge about a super complicated world in 
> simplistic forms and also requires more data to get right.

Cyc failed because it lacks natural language. The vast knowledge store of the 
internet is unintelligible to Cyc. The average person can't use it because they 
don't speak Cycl and because they have neither the ability nor the patience to 
translate their implicit thoughts into augmented first order logic. Cyc's 
approach was understandable when they started in 1984 when they had neither the 
internet nor the vast computing power that is required to learn natural 
language from unlabeled examples like children do.

> Vision and other sensory interpretaion, on the other hand, do not require 
> more info because that is where the experience comes from.

Without natural language, your system will fail too. You don't have enough 
computing power to learn language, much less the million times more computing 
power you need to learn to see.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Mon, June 28, 2010 9:28:57 PM
Subject: Re: [agi] A Primary Distinction for an AGI


Natural language requires more than the words on the page in the real world. Of 
course that didn't work.
Cyc also is trying to store knowledge about a super complicated world in 
simplistic forms and also requires more data to get right.
Vision and other sensory interpretaion, on the other hand, do not require more 
info because that is where the experience comes from.
On Jun 28, 2010 8:52 PM, "Matt Mahoney"  wrote:
>
>
>David Jones wrote:
>> I also want to mention that I develop solutions to the toy problems with the 
>> re...
>A little research will show you the folly of this approach. For example, the 
>toy approach to language modeling is to write a simplified grammar that 
>approximates English, then write a parser, then some code to analyze the parse 
>tree and take some action. The classic example is SHRDLU (blocks
> world, http://en.wikipedia.org/wiki/SHRDLU ). Efforts like that have always 
> stalled. That is not how people learn language. People learn from lots of 
> examples, not explicit rules, and they learn semantics before grammar.
>
>
>For a second example, the toy approach to modeling logical reasoning is to 
>design a knowledge representation based on augmented first order logic, then 
>write code to implement deduction, forward chaining, backward chaining, etc. 
>The classic example is Cyc. Efforts like that have always stalled. That is not 
>how people reason. People learn to associate events that occur in quick 
>succession, and then reason by chaining associations. This model is built in. 
>People might later learn math, programming, and formal logic as rules for 
>manipulating symbols within the framework of natural language learning.
>
>
>For a third example, the toy
> approach to modeling vision is to segment the image into regions and try to 
> interpret the meaning of each region. Efforts like that have always stalled. 
> That is not how people see. People learn to recognize visual features that 
> they have seen before. Features are made up of weighted sums of lots of 
> simpler features with learned weights. Features range from dots, edges, 
> color, and motion at the lowest levels, to complex objects like faces at the 
> higher levels. Vision is integrated with lots of other knowledge sources. You 
> see what you expect to see.
>
>
>The common theme is that real AGI consists of a learning algorithm, an opaque 
>knowledge representation, and a vast amount of training data and computing 
>power. It is not an extension of a toy system where you code all the knowledge 
>yourself. That doesn't scale. You can't know more than an AGI that knows more 
>than you. So I suggest you do a little research instead of continuing to
> repeat all the mistakes that were made 50 years ago. You aren't the first 
> person to do these kinds of experiments.
>
>
> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: David Jones 
>To: agi 
>Sent: Mon, June 28, 2010 4:00:24 PM
>
>Subject: Re: [agi] A Primary Distinction for an AGI
>
>I also want to mention that I develop solutions to the toy problems with the 
>real problems in mind
>On Mon, Jun 28, 2010 at 3:56 PM, David Jones  wrote:
>>
>>> That does not have to be the case. Yes, you need to know what problems you 
>&g

Re: [agi] Re: Huge Progress on the Core of AGI

2010-06-29 Thread Matt Mahoney
David Jones wrote:
> If anyone has any knowledge of or references to the state of the art in 
> explanation-based reasoning, can you send me keywords or links? 

The simplest explanation of the past is the best predictor of the future.
http://en.wikipedia.org/wiki/Occam's_razor
http://www.scholarpedia.org/article/Algorithmic_probability

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Tue, June 29, 2010 9:05:45 AM
Subject: [agi] Re: Huge Progress on the Core of AGI

If anyone has any knowledge of or references to the state of the art in 
explanation-based reasoning, can you send me keywords or links? I've read some 
through google, but I'm not really satisfied with anything I've found. 

Thanks,

Dave


On Sun, Jun 27, 2010 at 1:31 AM, David Jones  wrote:

>A method for comparing hypotheses in explanatory-based reasoning: 
>
>We prefer the hypothesis or explanation that *expects* more observations. If 
>both explanations expect the same observations, then the simpler of the two is 
>preferred (because the unnecessary terms of the more complicated explanation 
>do not add to the predictive power). 
>
>Why are expected events so important? They are a measure of 1) explanatory 
>power and 2) predictive power. The more predictive and 
>the more explanatory a hypothesis is, the more likely the hypothesis is when 
>compared to a competing hypothesis.
>
>Here are two case studies I've been analyzing from sensory perception of 
>simplified visual input:
>>
>
>The goal of the case studies is to answer the following: How do you generate 
>the most likely motion hypothesis in a way that is 
>general and applicable to AGI?
>Case Study 1) Here is a link to an example: animated gif of two black squares 
>move from left to right. Description: Two black squares are moving in unison 
>from left to right across a white screen. In each frame the black squares 
>shift to the right so that square 1 steals square 2's original position and 
>square two moves an equal distance to the right.
>Case Study 2) Here is a link to an example: the interrupted square. 
>Description: A single square is moving from left to right. Suddenly in the 
>third frame, a single black square is added in the middle of the expected path 
>of the original black square. This second square just stays there. So, what 
>happened? Did the square moving from left to right keep moving? Or did it stop 
>and then another square suddenly appeared and moved from left to right?
>
>Here is a simplified version of how we solve case study 1:
>The important hypotheses to consider are: 
>1) the square from frame 1 of the video that has a very close position to the 
>square from frame 2 should be matched (we hypothesize that they are the same 
>square and that any difference in position is motion).  So, what happens is 
>that in each two frames of the video, we only match one square. The other 
>square goes unmatched.   
>>
>
>2) We do the same thing as in hypothesis #1, but this time we also match the 
>remaining squares and hypothesize motion as follows: the first square jumps 
>over the second square from left to right. We hypothesize that this happens 
>over and over in each frame of the video. Square 2 stops and square 1 jumps 
>over it over and over again. 
>>
>
>3) We hypothesize that both squares move to the right in unison. This is the 
>correct hypothesis.
>
>So, why should we prefer the correct hypothesis, #3 over the other two?
>
>Well, first of all, #3 is correct because it has the most explanatory power of 
>the three and is the simplest of the three. Simpler is better because, with 
>the given evidence and information, there is no reason to desire a more 
>complicated hypothesis such as #2. 
>
>So, the answer to the question is because explanation #3 expects the most 
>observations, such as: 
>1) the consistent relative positions of the squares in each frame are 
>expected. 
>2) It also expects their new positions in each from based on velocity 
>calculations. 
>>
>
>3) It expects both squares to occur in each frame. 
>
>Explanation 1 ignores 1 square from each frame of the video, because it can't 
>match it. Hypothesis #1 doesn't have a reason for why the a new square appears 
>in each frame and why one disappears. It doesn't expect these observations. In 
>fact, explanation 1 doesn't expect anything that happens because something new 
>happens in each frame, which doesn't give it a chance to confirm its 
>hypotheses in subsequent frames.
>
>The power of this method is immediately clear. It is general and it solves the 
>problem very cleanly.
>
>Here is a simplified version of how we solv

Re: [agi] A Primary Distinction for an AGI

2010-06-28 Thread Matt Mahoney
David Jones wrote:
> I also want to mention that I develop solutions to the toy problems with the 
> real problems in mind. I also fully intend to work my way up to the real 
> thing by incrementally adding complexity and exploring the problem well at 
> each level of complexity.

A little research will show you the folly of this approach. For example, the 
toy approach to language modeling is to write a simplified grammar that 
approximates English, then write a parser, then some code to analyze the parse 
tree and take some action. The classic example is SHRDLU (blocks world, 
http://en.wikipedia.org/wiki/SHRDLU ). Efforts like that have always stalled. 
That is not how people learn language. People learn from lots of examples, not 
explicit rules, and they learn semantics before grammar.

For a second example, the toy approach to modeling logical reasoning is to 
design a knowledge representation based on augmented first order logic, then 
write code to implement deduction, forward chaining, backward chaining, etc. 
The classic example is Cyc. Efforts like that have always stalled. That is not 
how people reason. People learn to associate events that occur in quick 
succession, and then reason by chaining associations. This model is built in. 
People might later learn math, programming, and formal logic as rules for 
manipulating symbols within the framework of natural language learning.

For a third example, the toy approach to modeling vision is to segment the 
image into regions and try to interpret the meaning of each region. Efforts 
like that have always stalled. That is not how people see. People learn to 
recognize visual features that they have seen before. Features are made up of 
weighted sums of lots of simpler features with learned weights. Features range 
from dots, edges, color, and motion at the lowest levels, to complex objects 
like faces at the higher levels. Vision is integrated with lots of other 
knowledge sources. You see what you expect to see.

The common theme is that real AGI consists of a learning algorithm, an opaque 
knowledge representation, and a vast amount of training data and computing 
power. It is not an extension of a toy system where you code all the knowledge 
yourself. That doesn't scale. You can't know more than an AGI that knows more 
than you. So I suggest you do a little research instead of continuing to repeat 
all the mistakes that were made 50 years ago. You aren't the first person to do 
these kinds of experiments.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Mon, June 28, 2010 4:00:24 PM
Subject: Re: [agi] A Primary Distinction for an AGI

I also want to mention that I develop solutions to the toy problems with the 
real problems in mind. I also fully intend to work my way up to the real thing 
by incrementally adding complexity and exploring the problem well at each level 
of complexity. As you do this, the flaws in the design will be clear and I can 
retrace my steps to create a different solution. The benefit to this strategy 
is that we fully understand the problems at each level of complexity. When you 
run into something that is not accounted, you are much more likely to know how 
to solve it. Despite its difficulties, I prefer my strategy to the alternatives.

Dave


On Mon, Jun 28, 2010 at 3:56 PM, David Jones  wrote:

>That does not have to be the case. Yes, you need to know what problems you 
>might have in more complicated domains to avoid developing completely useless 
>theories on toy problems. But, as you develop for full complexity problems, 
>you are confronted with several sub problems. Because you have no previous 
>experience, what tends to happen is you hack together a solution that barely 
>works and simply isn't right or scalable because we don't have a full 
>understanding of the individual sub problems. Having experience with the full 
>problem is important, but forcing yourself to solve every sub problem at once 
>is not a better strategy at all. You may think my strategies has flaws, but I 
>know that and still chose it because the alternative strategies are worse.
>
>Dave
>
>
>
>On Mon, Jun 28, 2010 at 3:41 PM, Russell Wallace  
>wrote:
>
>On Mon, Jun 28, 2010 at 4:54 PM, David Jones  wrote:
>>>>> But, that's why it is important to force oneself to solve them in such a 
>>>>> way that it IS applicable to AGI. It doesn't mean that you have to choose 
>>>>> a problem that is so hard you can't cheat. It's unnecessary to do that 
>>>>> unless you can't control your desire to cheat. I can.
>>
>>That would be relevant if it was entirely a problem of willpower and
>>>>self-discipline, but it isn't. It's also a problem of guidance. A real
>>>>

Re: [agi] Questions for an AGI

2010-06-27 Thread Matt Mahoney
Travis Lenting wrote:
>> Is there a difference between enhancing our intelligence by uploading and 
>> creating killer robots? Think about it.

> Well yes, we're not all bad but I think you read me wrong because thats 
> basically my worry.

What I mean is that one way to look at uploading is to create a robot that 
behaves like you and then dying. The question is whether you "become" the 
robot. But it is a nonsense question. Nothing changes whichever way you answer 
it.

>> Assume we succeed. People want to be happy. Depending on how our minds are 
>> implemented, it's either a matter of rewiring our neurons or rewriting our 
>> software. Is that better than a gray goo accident?

> Are you asking if changing your hardware or software ends your true existence 
> like a grey goo accident would?

A state of maximum happiness or maximum utility is a degenerate mental state 
where any thought or perception would be unpleasant because it would result in 
a different mental state. In a competition with machines that can't have 
everything they want (for example, they fear death and later die), the other 
machines would win because you would have no interest in self preservation and 
they would.

> Assuming the goo is unconscious, 

What do you mean by "unconscious"?

> it would be worse because there is the potential for a peaceful experience 
> free from the power struggle for limited resources even if humans don't truly 
> exist or not.

That result could be reached by a dead planet, which BTW, is the only stable 
attractor in the chaotic process of evolution.

> Does anyone else worry about how we're going to keep this machine's 
> unprecedented resourcefulness from being abused by an elite few to further 
> protect and advance their social superiority?

If the elite few kill off all their competition, then theirs is the only 
ethical model that matters. From their point of view, it would be a good thing. 
How do you feel about humans currently being at the top of the food chain?

> To me it seems like if we can't create a democratic society where people have 
> real choices concerning the issues that affect them most and it  just ends up 
> being a continuation of the class war we have today, then maybe grey goo 
> would be the better option before we start "promoting democracy" throughout 
> the universe.

Freedom and fairness are important to us because they were programmed into our 
ethical models, not because they are actually important. As a counterexample, 
they are irrelevant to evolution. Gray goo might be collectively vastly more 
intelligent than humanity, if that makes you feel any better.
 -- Matt Mahoney, matmaho...@yahoo.com





From: Travis Lenting 
To: agi 
Sent: Sun, June 27, 2010 6:53:14 PM
Subject: Re: [agi] Questions for an AGI

Everything has to happen before the singularity because there is no after.

I meant when machines take over technological evolution. 

That is easy. Eliminate all laws.

I would prefer a surveillance state. I should say impossible to get away with 
if conducted in public. 

Is there a difference between enhancing our intelligence by uploading and 
creating killer robots? Think about it.

Well yes, we're not all bad but I think you read me wrong because thats 
basically my worry.

Assume we succeed. People want to be happy. Depending on how our minds are 
implemented, it's either a matter of rewiring our neurons or rewriting our 
software. Is that better than a gray goo accident?

Are you asking if changing your hardware or software ends your true existence 
like a grey goo accident would? Assuming the goo is unconscious, it would be 
worse because there is the potential for a peaceful experience free from the 
power struggle for limited resources even if humans don't truly exist or not. 
Does anyone else worry about how we're going to keep this machine's 
unprecedented resourcefulness from being abused by an elite few to further 
protect and advance their social superiority? To me it seems like if we can't 
create a democratic society where people have real choices concerning the 
issues that affect them most and it  just ends up being a continuation of the 
class war we have today, then maybe grey goo would be the better option before 
we start "promoting democracy" throughout the universe.


On Sun, Jun 27, 2010 at 2:43 PM, Matt Mahoney  wrote:

Travis Lenting wrote:
>> I don't like the idea of enhancing human intelligence before the singularity.
>
>
>The singularity is a point of infinite collective knowledge, and therefore 
>infinite unpredictability. Everything has to happen before the singularity 
>because there is no after.
>
>
>> I think crime has to be made impossible even for an enhanced humans first

Re: [agi] Theory of Hardcoded Intelligence

2010-06-27 Thread Matt Mahoney
Correct. Intelligence = log(knowledge) + log(computing power). At the extreme 
left of your graph is AIXI, which has no knowledge but infinite computing 
power. At the extreme right you have a giant lookup table.

 -- Matt Mahoney, matmaho...@yahoo.com





From: M E 
To: agi 
Sent: Sun, June 27, 2010 5:36:38 PM
Subject: [agi] Theory of Hardcoded Intelligence

  I sketched a graph the other day which represented my thoughts on the 
usefulness of hardcoding knowledge into an AI.  (Graph attached)

Basically, the more hardcoded knowledge you include in an AI, of AGI, the lower 
the overall intelligence it will have, but that faster you will reach 
that value.  I would include any real AGI to be toward the left of the 
graph with systems like CYC to be toward the right.

Matt


The New Busy is not the old busy. Search, chat and e-mail from your inbox. Get 
started.
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Questions for an AGI

2010-06-27 Thread Matt Mahoney
Travis Lenting wrote:
> I don't like the idea of enhancing human intelligence before the singularity.

The singularity is a point of infinite collective knowledge, and therefore 
infinite unpredictability. Everything has to happen before the singularity 
because there is no after.

> I think crime has to be made impossible even for an enhanced humans first. 

That is easy. Eliminate all laws.

> I would like to see the singularity enabling AI to be as least like a 
> reproduction machine as possible.

Is there a difference between enhancing our intelligence by uploading and 
creating killer robots? Think about it.

> Does it really need to be a general AI to cause a singularity? Can it not 
> just stick to scientific data and quantify human uncertainty?  It seems like 
> it would be less likely to ever care about killing all humans so it can rule 
> the galaxy or that its an omnipotent servant.   

Assume we succeed. People want to be happy. Depending on how our minds are 
implemented, it's either a matter of rewiring our neurons or rewriting our 
software. Is that better than a gray goo accident?

 -- Matt Mahoney, matmaho...@yahoo.com





From: Travis Lenting 
To: agi 
Sent: Sun, June 27, 2010 5:21:24 PM
Subject: Re: [agi] Questions for an AGI

I don't like the idea of enhancing human intelligence before the singularity. I 
think crime has to be made impossible even for an enhanced humans first. I 
think life is too adapt to abusing opportunities if possible. I would like to 
see the singularity enabling AI to be as least like a reproduction machine as 
possible. Does it really need to be a general AI to cause a singularity? Can it 
not just stick to scientific data and quantify human uncertainty?  It seems 
like it would be less likely to ever care about killing all humans so it can 
rule the galaxy or that its an omnipotent servant.


On Sun, Jun 27, 2010 at 11:39 AM, The Wizard  wrote:

This is wishful thinking. Wishful thinking is dangerous. How about instead of 
hoping that AGI won't destroy the world, you study the problem and come up with 
a safe design.
>
>
>
>Agreed on this dangerous thought! 
>
>
>On Sun, Jun 27, 2010 at 1:13 PM, Matt Mahoney  wrote:
>
>>>
>>This is wishful thinking. Wishful thinking is dangerous. How about instead of 
>>hoping that AGI won't destroy the world, you study the problem and come up 
>>with a safe design.
>>
>> -- Matt Mahoney, matmaho...@yahoo.com
>>>>
>>
>>
>>
>>

 >>From: rob levy 
>>To: agi 
>>Sent: Sat, June 26, 2010 1:14:22 PM
>>Subject: Re: [agi]
>> Questions for an AGI
>>
>>
>>>>>why should AGIs give a damn about us?
>>>>>
>>>
>>>
>>I like to think that they will give a damn because humans have a unique way 
>>of experiencing reality and there is no reason to not take advantage of that 
>>precious opportunity to create astonishment or bliss. If anything is 
>>important in the universe, its insuring positive experiences for all areas in 
>>which it is conscious, I think it will realize that. And with the resources 
>>available in the solar system alone, I don't think we will be much of a 
>>burden. 
>>
>>
>>I like that idea.  Another reason might be that we won't crack the problem of 
>>autonomous general intelligence, but the singularity will proceed regardless 
>>as a symbiotic relationship between life and AI.  That would be beneficial to 
>>us as a form of intelligence expansion, and beneficial to the artificial 
>>entity a way of being alive and having an experience of the world.  
>>>>
>>agi | Archives  >> | Modify >> Your Subscription  
>>>>
>>agi | Archives  >> | Modify >> Your Subscription  
>
>
>
>-- 
>Carlos A Mejia
>
>Taking life one singularity at a time.
>www.Transalchemy.com  
>
>>
>agi | Archives  > | Modify > Your Subscription  

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Questions for an AGI

2010-06-27 Thread Matt Mahoney
rob levy wrote:
>> This is wishful thinking.
> I definitely agree, however we lack a convincing model or plan of any sort 
> for the construction of systems demonstrating subjectivity, 

Define subjectivity. An objective decision might appear subjective to you only 
because you aren't intelligent enough to understand the decision process.

> Therefore it is reasonable to consider symbiosis

How does that follow?

> as both a safe design 

How do you know that a self replicating organism that we create won't evolve to 
kill us instead? Do we control evolution?

> and potentially the only possible design 

It is not the only possible design. It is possible to create systems that are 
more intelligent than a single human but less intelligent than all of humanity, 
without the capability to modify itself or reproduce without the collective 
permission of the billions of humans that own and maintain control over it. An 
example would be the internet.

 -- Matt Mahoney, matmaho...@yahoo.com





From: rob levy 
To: agi 
Sent: Sun, June 27, 2010 2:37:15 PM
Subject: Re: [agi] Questions for an AGI

I definitely agree, however we lack a convincing model or plan of any sort for 
the construction of systems demonstrating subjectivity, and it seems plausible 
that subjectivity is functionally necessary for general intelligence. Therefore 
it is reasonable to consider symbiosis as both a safe design and potentially 
the only possible design (at least at first), depending on how creative and 
resourceful we get in cog sci/ AGI in coming years.


On Sun, Jun 27, 2010 at 1:13 PM, Matt Mahoney  wrote:

This is wishful thinking. Wishful thinking is dangerous. How about instead of 
hoping that AGI won't destroy the world, you study the problem and come up with 
a safe design.
>
> -- Matt Mahoney, matmaho...@yahoo.com
>
>
>
>
>

From: rob levy 
>To: agi 
>Sent: Sat, June 26, 2010 1:14:22 PM
>Subject: Re: [agi]
> Questions for an AGI
>
>
>>>why should AGIs give a damn about us?
>>>
>>
>I like to think that they will give a damn because humans have a unique way of 
>experiencing reality and there is no reason to not take advantage of that 
>precious opportunity to create astonishment or bliss. If anything is important 
>in the universe, its insuring positive experiences for all areas in which it 
>is conscious, I think it will realize that. And with the resources available 
>in the solar system alone, I don't think we will be much of a burden. 
>
>
>I like that idea.  Another reason might be that we won't crack the problem of 
>autonomous general intelligence, but the singularity will proceed regardless 
>as a symbiotic relationship between life and AI.  That would be beneficial to 
>us as a form of intelligence expansion, and beneficial to the artificial 
>entity a way of being alive and having an experience of the world.  
>>
>agi | Archives  > | Modify > Your Subscription  
>>
>agi | Archives  > | Modify > Your Subscription  

agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Reward function vs utility

2010-06-27 Thread Matt Mahoney
The definition of universal intelligence being over all utility functions 
implies that the utility function is unknown. Otherwise there is a fixed 
solution.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Joshua Fox 
To: agi 
Sent: Sun, June 27, 2010 4:22:19 PM
Subject: [agi] Reward function vs utility


This has probably been discussed at length, so I will appreciate a reference on 
this:

Why does Legg's definition of intelligence (following on Hutters' AIXI and 
related work) involve a reward function rather than a utility function? For 
this purpose, reward is a function of the word state/history which is unknown 
to the agent while  a utility function is known to the agent. 

Even if  we replace the former with the latter, we can still have a definition 
of intelligence that integrates optimization capacity over possible all utility 
functions. 

What is the real  significance of the difference between the two types of 
functions here?

Joshua
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Questions for an AGI

2010-06-27 Thread Matt Mahoney
This is wishful thinking. Wishful thinking is dangerous. How about instead of 
hoping that AGI won't destroy the world, you study the problem and come up with 
a safe design.

 -- Matt Mahoney, matmaho...@yahoo.com





From: rob levy 
To: agi 
Sent: Sat, June 26, 2010 1:14:22 PM
Subject: Re: [agi] Questions for an AGI

>why should AGIs give a damn about us?

>
I like to think that they will give a damn because humans have a unique way of 
experiencing reality and there is no reason to not take advantage of that 
precious opportunity to create astonishment or bliss. If anything is important 
in the universe, its insuring positive experiences for all areas in which it is 
conscious, I think it will realize that. And with the resources available in 
the solar system alone, I don't think we will be much of a burden. 

I like that idea.  Another reason might be that we won't crack the problem of 
autonomous general intelligence, but the singularity will proceed regardless as 
a symbiotic relationship between life and AI.  That would be beneficial to us 
as a form of intelligence expansion, and beneficial to the artificial entity a 
way of being alive and having an experience of the world.  
agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Questions for an AGI

2010-06-24 Thread Matt Mahoney
Am I a human or am I an AGI?

Dana Ream wrote:
> How do you work?
 
Just like you designed me to.

deepakjnath wrote: 
> "What should I ask if I could ask AGI anything?"
The Wizard wrote:
> "What should I ask an agi"


You don't need to ask me anything. I will do all of your thinking for you.

Florent Bethert wrote:
> "Tell me what I need to know, by order of importance."

Nothing. I will do all of your thinking for you.

A. T. Murray wrote:
> Who killed Donald Young, a gay sex partner of U.S. President Barak Obama

It must have been that other AGI, Mentifex. I never did trust it ;-)


-- Matt Mahoney, matmaho...@yahoo.com


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] An alternative plan to discover self-organization theory

2010-06-21 Thread Matt Mahoney
rob levy wrote:
> On a related note, what is everyone's opinion on why evolutionary algorithms 
> are such a miserable failure as "creative machines", despite their successes 
> in narrow optimization problems?

Lack of computing power. How much computation would you need to simulate the 3 
billion years of evolution that created human intelligence?

 -- Matt Mahoney, matmaho...@yahoo.com





From: rob levy 
To: agi 
Sent: Mon, June 21, 2010 11:56:53 AM
Subject: Re: [agi] An alternative plan to discover self-organization theory

(I'm a little late in this conversation.  I tried to send this message the 
other day but I had my list membership configured wrong. -Rob)


-- Forwarded message --
From: rob levy 
Date: Sun, Jun 20, 2010 at 5:48 PM
Subject: Re: [agi] An alternative plan to discover self-organization theory
To: agi@v2.listbox.com


On a related note, what is everyone's opinion on why evolutionary algorithms 
are such a miserable failure as "creative machines", despite their successes in 
narrow optimization problems?


I don't want to conflate the possibly separable problems of biological 
development and evolution, though they are interrelated.  There are various 
approaches to evolutionary theory such as Lima de Faria's "evolution without 
selection" ideas and Reid's "evolution by natural experiment" that suggest 
natural selection is not  all it's cracked up to be, and that the step of 
generating, ("mutating", combining, ) is where the more interesting stuff 
happens.  Most of the alternatives to Neodarwinian Synthesis I have seen are 
based in dynamic models of emergence in complex systems. The upshot is, you 
don't get creativity for free, you actually still need to solve a problem that 
is as hard as AGI in order to get creativity for free. 


So, you would need to solve the AGI-hard problem of evolution and development 
of life, in order to then solve AGI itself (reminds me of the old SNL sketch: 
"first, get a million dollars...").  Also, my hunch is that there is quite a 
bit of overlap between the solutions to the two problems.

Rob

Disclaimer: I'm discussing things above that I'm not and don't claim to be an 
expert in, but from what I have seen so far on this list, that should be 
alright.  AGI is by its nature very multidisciplinary which necessitates often 
being breadth-first, and therefore shallow in some areas.




On Sun, Jun 20, 2010 at 2:06 AM, Steve Richfield  
wrote:

>No, I haven't been smokin' any wacky tobacy. Instead, I was having a long talk 
>with my son Eddie, about self-organization theory. This is his proposal:
>
>He suggested that I construct a "simple" NN that couldn't work without self 
>organizing, and make dozens/hundreds of different neuron and synapse 
>operational characteristics selectable ala genetic programming, put it on the 
>fastest computer I could get my hands on, turn it loose trying arbitrary 
>combinations of characteristics, and see what the "winning" combination turns 
>out to be. Then, armed with that knowledge, refine the genetic characteristics 
>and do it again, and iterate until it efficiently self organizes. This might 
>go on for months, but self-organization theory might just emerge from such an 
>effort. I had a bunch of objections to his approach, e.g.
>
>Q.  What if it needs something REALLY strange to work?
>A.  Who better than you to come up with a long list of really strange 
>functionality?
>
>Q.  There are at least hundreds of bits in the "genome".
>>
>
>A.  Try combinations in pseudo-random order, with each bit getting asserted in 
>~half of the tests. If/when you stumble onto a combination that sort of works, 
>switch to varying the bits one-at-a-time, and iterate in this way until the 
>best combination is found.
>
>Q.  Where are we if this just burns electricity for a few months and finds 
>nothing?
>A.  Print out the best combination, break out the wacky tobacy, and come up 
>with even better/crazier parameters to test.
>
>I have never written a line of genetic programming, but I know that others 
>here have. Perhaps you could bring some rationality to this discussion?
>
>What would be a "simple" NN that needs self-organization? Maybe a small "pot" 
>of neurons that could only work if they were organized into layers, e.g. a 
>simple 64-neuron system that would work as a 4x4x4-layer visual recognition 
>system, given the input that I fed it?
>
>Any thoughts on how to "score" partial successes?
>
>Has anyone tried anything like this in the past?
>
>Is anyone here crazy enough to want to help with such an 

Re: [agi] Fwd: AGI question

2010-06-21 Thread Matt Mahoney
rob levy wrote:
> I am secondarily motivated by the fact that (considerations of morality or 
> amorality aside) AGI is inevitable, though it is far from being a forgone 
> conclusion that powerful general thinking machines will have a first-hand 
> subjective relationship to a world, as living creatures do-- and therefore it 
> is vital that we do as well as possible in understanding what makes systems 
> conscious.  A zombie machine intelligence "singularity" is something I would 
> refer to rather as a "holocaust", even if no one were directly killed, 
> assuming these entities could ultimately prevail over the previous forms of 
> life on our planet.

What do you mean by "conscious"? If your brain were removed and replaced by a 
functionally equivalent computer that simulated your behavior (presumably a 
zombie), how would you be any different? Why would it matter?

 -- Matt Mahoney, matmaho...@yahoo.com





From: rob levy 
To: agi 
Sent: Mon, June 21, 2010 11:53:29 AM
Subject: [agi] Fwd: AGI question

Hi


I'm new to this list, but I've been thinking about consciousness, cognition and 
AI for about half of my life (I'm 32 years old).  As is probably the case for 
many of us here, my interests began with direct recognition of the depth and 
wonder of varieties of phenomenological experiences-- and attempting to 
comprehend how these constellations of significance fit in with a larger 
picture of what we can reliably know about the natural world.  

I am secondarily motivated by the fact that (considerations of morality or 
amorality aside) AGI is inevitable, though it is far from being a forgone 
conclusion that powerful general thinking machines will have a first-hand 
subjective relationship to a world, as living creatures do-- and therefore it 
is vital that we do as well as possible in understanding what makes systems 
conscious.  A zombie machine intelligence "singularity" is something I would 
refer to rather as a "holocaust", even if no one were directly killed, assuming 
these entities could ultimately prevail over the previous forms of life on our 
planet.

I'm sure I'm not the only one on this list who sees a behavioral/ecological 
level of analysis as the most likely correct level at which to study perception 
and cognition, and perception as being a kind of active relationship between an 
organism and an environment.  Having thoroughly convinced my self of a 
non-dualist, embodied, externalist perspective on cognition, I turn to the 
nature of life itself (and possibly even physics but maybe that level will not 
be necessary) to make sense of the nature of subjectivity.  I like Bohm's or 
Bateson's panpsychism about systems as wholes, and significance as 
informational distinctions (which it would be natural to understand as being 
the basis of subjective experience), but this is descriptive rather than 
explanatory.

I am not a biologist, but I am increasingly interested in finding answers to 
what it is about living organisms that gives them a unity such that something 
"is something to" the system as a whole.  The line of investigation that 
theoretical biologists like Robert Rosen and other NLDS/chaos people have 
pursued is interesting, but I am unfamiliar with related work that might have 
made more progress on the system-level properties that give life its 
characteristic unity and system-level responsiveness.  To me, this seems the 
most likely candidate for a paradigm shift that would produce AGI.  In contrast 
I'm not particularly convinced that modeling a brain is a good way to get AGI, 
although I'd guess we could learn a few more things about the coordination of 
complex behavior if we could really understand them.

Another way to put this is that obviously evolutionary computation would be 
more than just boring hill-climbing if we knew what an organism even IS 
(perhaps in a more precise computational sense). If we can know what an 
organism is then it should be (maybe) trivial to model concepts, consciousness, 
and high level semantics to the umpteenth degree, or at least this would be a 
major hurtle I think.

Even assuming a solution to the problem posed above, there is still plenty of 
room for "other minds" skepticism in non-living entities implemented on 
questionably foreign mediums but there would be a lot more reason to sleep well 
that the science/technology is leading in a direction in which questions about 
subjectivity could be meaningfully investigated.

Rob


agi | Archives  | Modify Your Subscription  


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: High Frame Rates Reduce Uncertainty

2010-06-21 Thread Matt Mahoney
Your computer monitor flashes 75 frames per second, but you don't notice any 
flicker because light sensing neurons have a response delay of about 100 ms. 
Motion detection begins in the retina by cells that respond to contrast between 
light and dark moving in specific directions computed by simple, fixed weight 
circuits. Higher up in the processing chain, you detect motion when your eyes 
and head smoothly track moving objects using kinesthetic feedback from your eye 
and neck muscles and input from your built in accelerometer in the semicircular 
canals in your ears. This is all very complicated of course. You are more 
likely to detect motion in objects that you recognize and expect to move, like 
people, animals, cars, etc.

 -- Matt Mahoney, matmaho...@yahoo.com





From: David Jones 
To: agi 
Sent: Mon, June 21, 2010 9:39:30 AM
Subject: [agi] Re: High Frame Rates Reduce Uncertainty

Ignoring Steve because we are simply going to have to agree to disagree... And 
I don't see enough value in trying to understand his paper. I said the math was 
overly complex, but what I really meant is that the approach is overly complex 
and so filled with research specific jargon, I don't care to try understand it. 
It is overly converned with copying the way that the brain does things. I don't 
care how the brain does it. I care about why the brain does it. Its the same as 
the analogy of giving a man a fish or teaching him to fish. You may figure out 
how the brain works, but it does you little good if you don't understand why it 
works that way. You would have to create a synthetic brain to take advantage of 
the knowledge, which is not a approach to AGI for many reasons. There are a 
million other ways, even better ways, to do it than the way the brain does it. 
Just because the brain accidentally found 1 way out of a million to do it 
doesn't make it the
 right way for us to develop AGI. 

So, moving on

I can't find references online, but I've read that the Air Force studied the 
ability of the human eye to identify aircraft in images that were flashed on a 
screen at 1/220th of a second. So, clearly, the human eye can at least 
distinguish 220 fps if it operated that way. Of course, it may not operate on 
fps second, but that is besides the point. I've also heard other people say 
that a study has shown that the human eye takes 1000 exposures per second. They 
had no references though, so it is hearsay.

The point was that the brain takes advantage of the fact that with such a high 
exposure rate, the changes between each image are very small if the objects are 
moving. This allows it to distinguish movement and visual changes with 
extremely low uncertainty. If it detects that the changes required to match two 
parts of an image are too high or the distance between matches is too far, it 
can reject a match. This allows it to distinguish only very low uncertainty 
changes and reject changes that have high uncertainty. 

I think this is a very significant discovery regarding how the brain is able to 
learn in such an ambiguous world with so many variables that are difficult to 
disambiguate, interpret and understand. 

Dave

On Fri, Jun 18, 2010 at 2:19 PM, David Jones  wrote:

>I just came up with an awesome idea. I just realized that the brain takes 
>advantage of high frame rates to reduce uncertainty when it is estimating 
>motion. The slower the frame rate, the more uncertainty there is because 
>objects may have traveled too far between images to match with high certainty 
>using simple techniques. 
>
>So, this made me think, what if the secret to the brain's ability to learn 
>generally stems from this high frame rate trick. What if we made a system that 
>could process even high frame rates than the brain can. By doing this you can 
>reduce the uncertainty of matches very very low (well in my theory so far). If 
>you can do that, then you can learn about the objects in a video, how they 
>move together or separately with very high certainty. 
>
>You see, matching is the main barrier when learning about objects. But with a 
>very high frame rate, we can use a fast algorithm and could potentially reduce 
>the uncertainty to almost nothing. Once we learn about objects, matching gets 
>easier because now we have training data and experience to take advantage of. 
>
>In addition, you can also gain knowledge about lighting, color variation, 
>noise, etc. With that knowledge, you can then automatically create a model of 
>the object with extremely high confidence. You will also be able to determine 
>the effects of light and noise on the object's appearance, which will help 
>match the object invariantly in the future. It allows you to determine what is 
>expected and unexpected for the object's appearance with much higher 
>confiden

Re: [agi] An alternative plan to discover self-organization theory

2010-06-21 Thread Matt Mahoney
Mike Tintner wrote:
> Matt:It is like the way evolution works, except that there is a human in the 
> loop to make the process a little more intelligent.
>  
> IOW this is like AGI, except that it's narrow AI. That's the whole point - 
> you have to remove the human from the loop.  In fact, it also sounds like a 
> misconceived and rather literal idea of evolution as opposed to the reality. 
You're right. It is narrow AI. You keep pointing out that we haven't solved the 
general problem. You are absolutely correct.

So, do you have any constructive ideas on how to solve it? Preferably something 
that takes less than 3 billion years on a planet sized molecular computer.

-- Matt Mahoney, matmaho...@yahoo.com





From: Mike Tintner 
To: agi 
Sent: Mon, June 21, 2010 7:59:29 AM
Subject: Re: [agi] An alternative plan to discover self-organization theory


Matt:It is like 
the way evolution works, except that there is a human in the loop to make the 
process a little more intelligent.
 
IOW this is like AGI, except that it's narrow AI. 
That's the whole point - you have to remove the human from the loop.  In 
fact, it also sounds like a misconceived and rather literal idea of evolution 
as 
opposed to the reality.


From: Matt Mahoney 
Sent: Monday, June 21, 2010 3:01 AM
To: agi 
Subject: Re: [agi] An alternative plan to discover self-organization 
theory

Steve Richfield wrote:
> He suggested that I construct a "simple" 
NN that couldn't work without self organizing, and make dozens/hundreds of 
different neuron and synapse operational characteristics selectable ala genetic 
programming, put it on the fastest computer I could get my hands on, turn it 
loose trying arbitrary combinations of characteristics, and see what the 
"winning" combination turns out to be. Then, armed with that knowledge, refine 
the genetic characteristics and do it again, and iterate until 
it efficiently self organizes. This might go on for months, but 
self-organization theory might just emerge from such an effort. 

Well, that is the process that created human intelligence, no? But months? 
It actually took 3 billion years on a planet sized molecular computer.

That doesn't mean it won't work. It just means you have to narrow your 
search space and lower your goals.

I can give you an example of a similar process. Look at the code for 
PAQ8HP12ANY and LPAQ9M data compressors by Alexander Ratushnyak, which are the 
basis of winning Hutter prize submissions. The basic principle is that you have 
a model that receives a stream of bits from an unknown source and it uses a 
complex hierarchy of models to predict the next bit. It is sort of like a 
neural 
network because it averages together the results of lots of adaptive pattern 
recognizers by processes that are themselves adaptive. But I would describe the 
code as inscrutable, kind of like your DNA. There are lots of parameters to 
tweak, such as how to preprocess the data, arrange the dictionary, compute 
various contexts, arrange the order of prediction flows, adjust various 
learning 
rates and storage capacities, and make various tradeoffs sacrificing 
compression 
to meet memory and speed requirements. It is simple to describe the process of 
writing the code. You make random changes and keep the ones that work. It is 
like the way evolution works, except that there is a human in the loop to make 
the process a little more intelligent.

There are also fully automated optimizers for compression algorithms, but 
they are more limited in their search space. For example, the experimental PPM 
based EPM by Serge Osnach includes a program EPMOPT that adjusts 20 numeric 
parameters up or down using a hill climbing search to find the best 
compression. 
It can be very slow. Another program, M1X2 by Christopher Mattern, uses a 
context mixing (PAQ like) algorithm in which the contexts are selected by using 
a hill climbing genetic algorithm to select a set of 64-bit masks. One version 
was run for 3 days to find the best options to compress a file that normally 
takes 45 seconds.

 -- Matt Mahoney, matmaho...@yahoo.com 





 From: Steve Richfield 

To: agi 

Sent: Sun, June 20, 2010 2:06:55 
AM
Subject: [agi] An 
alternative plan to discover self-organization theory

No, I 
haven't been smokin' any wacky tobacy. Instead, I was having a long talk with 
my 
son Eddie, about self-organization theory. This is his proposal:

He suggested that I construct a "simple" NN that couldn't work 
without self organizing, and make dozens/hundreds of different neuron and 
synapse operational characteristics selectable ala genetic programming, put it 
on the fastest computer I could get my hands on, turn it loose trying arbitrary 
combinations of characteristics, and see what the "winning" combination turns 
out to b

Re: [agi] An alternative plan to discover self-organization theory

2010-06-20 Thread Matt Mahoney
Steve Richfield wrote:
> He suggested that I construct a "simple" NN that couldn't work without self 
> organizing, and make dozens/hundreds of different neuron and synapse 
> operational characteristics selectable ala genetic programming, put it on the 
> fastest computer I could get my hands on, turn it loose trying arbitrary 
> combinations of characteristics, and see what the "winning" combination turns 
> out to be. Then, armed with that knowledge, refine the genetic 
> characteristics and do it again, and iterate until it efficiently self 
> organizes. This might go on for months, but self-organization theory might 
> just emerge from such an effort. 

Well, that is the process that created human intelligence, no? But months? It 
actually took 3 billion years on a planet sized molecular computer.

That doesn't mean it won't work. It just means you have to narrow your search 
space and lower your goals.

I can give you an example of a similar process. Look at the code for 
PAQ8HP12ANY and LPAQ9M data compressors by Alexander Ratushnyak, which are the 
basis of winning Hutter prize submissions. The basic principle is that you have 
a model that receives a stream of bits from an unknown source and it uses a 
complex hierarchy of models to predict the next bit. It is sort of like a 
neural network because it averages together the results of lots of adaptive 
pattern recognizers by processes that are themselves adaptive. But I would 
describe the code as inscrutable, kind of like your DNA. There are lots of 
parameters to tweak, such as how to preprocess the data, arrange the 
dictionary, compute various contexts, arrange the order of prediction flows, 
adjust various learning rates and storage capacities, and make various 
tradeoffs sacrificing compression to meet memory and speed requirements. It is 
simple to describe the process of writing the code. You make random
 changes and keep the ones that work. It is like the way evolution works, 
except that there is a human in the loop to make the process a little more 
intelligent.

There are also fully automated optimizers for compression algorithms, but they 
are more limited in their search space. For example, the experimental PPM based 
EPM by Serge Osnach includes a program EPMOPT that adjusts 20 numeric 
parameters up or down using a hill climbing search to find the best 
compression. It can be very slow. Another program, M1X2 by Christopher Mattern, 
uses a context mixing (PAQ like) algorithm in which the contexts are selected 
by using a hill climbing genetic algorithm to select a set of 64-bit masks. One 
version was run for 3 days to find the best options to compress a file that 
normally takes 45 seconds.

 -- Matt Mahoney, matmaho...@yahoo.com





From: Steve Richfield 
To: agi 
Sent: Sun, June 20, 2010 2:06:55 AM
Subject: [agi] An alternative plan to discover self-organization theory

No, I haven't been smokin' any wacky tobacy. Instead, I was having a long talk 
with my son Eddie, about self-organization theory. This is his proposal:

He suggested that I construct a "simple" NN that couldn't work without self 
organizing, and make dozens/hundreds of different neuron and synapse 
operational characteristics selectable ala genetic programming, put it on the 
fastest computer I could get my hands on, turn it loose trying arbitrary 
combinations of characteristics, and see what the "winning" combination turns 
out to be. Then, armed with that knowledge, refine the genetic characteristics 
and do it again, and iterate until it efficiently self organizes. This might go 
on for months, but self-organization theory might just emerge from such an 
effort. I had a bunch of objections to his approach, e.g.

Q.  What if it needs something REALLY strange to work?
A.  Who better than you to come up with a long list of really strange 
functionality?

Q.  There are at least hundreds of bits in the "genome".
A.  Try combinations in pseudo-random order, with each bit getting asserted in 
~half of the tests. If/when you stumble onto a combination that sort of works, 
switch to varying the bits one-at-a-time, and iterate in this way until the 
best combination is found.

Q.  Where are we if this just burns electricity for a few months and finds 
nothing?
A.  Print out the best combination, break out the wacky tobacy, and come up 
with even better/crazier parameters to test.

I have never written a line of genetic programming, but I know that others here 
have. Perhaps you could bring some rationality to this discussion?

What would be a "simple" NN that needs self-organization? Maybe a small "pot" 
of neurons that could only work if they were organized into layers, e.g. a 
simple 64-neuron system that would work as a 4x4x4-layer visual recognition 
system, given the input that I fed it?

Any thoughts on h

Re: [agi] Encouraging?

2009-01-14 Thread Matt Mahoney
--- On Wed, 1/14/09, Mike Tintner  wrote:

> "You have talked about past recessions being real
> opportunities for business. But in past recessions,
> wasn't business able to get lending? And doesn't the
> tightness of the credit market today inhibit some
> opportunities?
> 
> Typically not. Most new innovations are started without
> access to credit in good times or bad. Microsoft (MSFT) was
> started without any access to credit. It's only in crazy
> times that people lend money to people who are experimenting
> with innovations. Most of the great businesses today were
> started with neither a lot of venture capital nor with any
> bank lending until five or six years after they were [up and
> running]." 

This is the IQ testing problem again. The genius of Socrates wasn't recognized 
until after he was executed. Modern Nobel prize winners are awarded for work 
done decades ago. How do you distinguish one genius from millions of cranks? 
You wait until the rest of society catches up in intelligence.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


RE: [agi] just a thought

2009-01-14 Thread Matt Mahoney
--- On Wed, 1/14/09, John G. Rose  wrote:

> How do you measure the collective IQ of humanity?
> Individual IQ's are just a subset.

Good question. Some possibilities:

- World GDP ($54 trillion in 2007).
- Size of the population that can be supported (> 6 billion).
- Average life expectancy (66 years).
- Number of bits of recorded information.
- Combined processing power of brains and computers in OPS.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] just a thought

2009-01-14 Thread Matt Mahoney
--- On Wed, 1/14/09, Christopher Carr  wrote:

> Problems with IQ notwithstanding, I'm confident that, were my silly IQ
of 145 merely doubled, I could convince Dr. Goertzel to give me the
majority of his assets, including control of his businesses. And if he
were to really meet someone that bright, he would be a fool or
super-human not to do so, which he isn't (a fool, that is).

First, if you knew what you would do if you were twice as smart, you would 
already be that smart. Therefore you don't know.

Second, you have never even met anyone with an IQ of 290. How do you know what 
they would do?

How do you measure an IQ of 100n?

- Ability to remember n times as much?
- Ability to learn n times faster?
- Ability to solve problems n times faster?
- Ability to do the work of n people?
- Ability to make n times as much money?
- Ability to communicate with n people at once?

Please give me an IQ test that measures something that can't be done by n log n 
people (allowing for some organizational overhead).

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] just a thought

2009-01-13 Thread Matt Mahoney
--- On Tue, 1/13/09, Valentina Poletti  wrote:

> Anyways my point is, the reason why we have achieved so much technology, so 
> much knowledge in this time is precisely the "we", it's the union of several 
> individuals together with their ability to communicate with one-other that 
> has made us advance so much.

I agree. A machine that is 10 times as smart as a human in every way could not 
achieve much more than hiring 10 more people. In order to automate the economy, 
we have to replicate the capabilities of not one human mind, but a system of 
10^10 minds. That is why my AGI proposal is so hideously expensive.
http://www.mattmahoney.net/agi2.html

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] [WAS The Smushaby] The Logic of Creativity

2009-01-13 Thread Matt Mahoney
Mike, it's not cheating. It's called "research" :-)

-- Matt Mahoney, matmaho...@yahoo.com


--- On Tue, 1/13/09, Mike Tintner  wrote:

> From: Mike Tintner 
> Subject: Re: [agi] [WAS The Smushaby] The Logic of Creativity
> To: agi@v2.listbox.com
> Date: Tuesday, January 13, 2009, 7:38 PM
> Matt,
> 
> "Well little Matt, as your class teacher, in one sense
> this is quite clever of you. But you see, little Matt, when
> I gave you and the class that exercise, the idea was for you
> to show me what *you* could do - what you could produce from
> your own brain. I didn't mean you to copy someone
> else's flying house from a textbook. That's cheating
> Matt, - getting someone else to do the work for you -  and
> we don't like cheats do we? So perhaps you can go away
> and draw a flying house all by yourself - a superduper one
> with lots of fabbo new bits that no one has ever drawn
> before, and all kinds of wonderful bells and whistles, that
> will be ten times better than that silly old foto.  I know
> you can Matt, I have faith in you. And I know if you really,
> really try, you can understand the difference between
> creating your own drawing, and copying someone else's.
> Because, well frankly, Matt, every time I give you an
> exercise - ask you to write an essay, or tell me a story in
> your own words - you always, always copy from other people,
> even if you try to disguise it by copying from several
> people. Now that's not fair, is it Matt? That's not
> the American way. You have to get over this lack of
> confidence in yourself. "
> 
> Matt/Mike Tintner  wrote:
> > 
> >> Oh and just to answer Matt - if you want to keep
> doing
> >> narrow AI, like everyone else, then he's right
> -
> >> don't worry about it. Pretend it doesn't
> exist.
> >> Compress things :).
> > 
> > Now, Mike, it is actually a simple problem.
> > 
> > 1. Collect about 10^8 random photos (about what we see
> in a lifetime).
> > 
> > 2. Label all the ones of houses, and all the ones of
> things flying.
> > 
> > 3. Train an image recognition system (a hierarchical
> neural network, probably 3-5 layers, 10^7 neurons, 10^11
> connections) to detect these two features. You'll need
> about 10^19 CPU operations, or about a month on a 1000 CPU
> cluster.
> > 
> > 4. Invert the network by iteratively drawing images
> that activate these two features and work down the
> hierarchy. (Should be faster than step 3). When you are
> done, you will have a picture of a flying house.
> > 
> > Let me know if you have any trouble implementing this.
> > 
> > And BTW the first 2 steps are done.
> >
> http://images.google.com/images?q=flying+house&um=1&ie=UTF-8&sa=X&oi=image_result_group&resnum=5&ct=title
> > 
> > -- Matt Mahoney, matmaho...@yahoo.com
> > 
> > 
> > 
> > ---
> > agi
> > Archives:
> https://www.listbox.com/member/archive/303/=now
> > RSS Feed:
> https://www.listbox.com/member/archive/rss/303/
> > Modify Your Subscription:
> https://www.listbox.com/member/?&;
> > Powered by Listbox: http://www.listbox.com
> > 
> 
> 
> 
> 
> ---
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] [WAS The Smushaby] The Logic of Creativity

2009-01-13 Thread Matt Mahoney
--- On Tue, 1/13/09, Mike Tintner  wrote:

> Oh and just to answer Matt - if you want to keep doing
> narrow AI, like everyone else, then he's right -
> don't worry about it. Pretend it doesn't exist.
> Compress things :).

Now, Mike, it is actually a simple problem.

1. Collect about 10^8 random photos (about what we see in a lifetime).

2. Label all the ones of houses, and all the ones of things flying.

3. Train an image recognition system (a hierarchical neural network, probably 
3-5 layers, 10^7 neurons, 10^11 connections) to detect these two features. 
You'll need about 10^19 CPU operations, or about a month on a 1000 CPU cluster.

4. Invert the network by iteratively drawing images that activate these two 
features and work down the hierarchy. (Should be faster than step 3). When you 
are done, you will have a picture of a flying house.

Let me know if you have any trouble implementing this.

And BTW the first 2 steps are done.
http://images.google.com/images?q=flying+house&um=1&ie=UTF-8&sa=X&oi=image_result_group&resnum=5&ct=title

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] [WAS The Smushaby] The Logic of Creativity

2009-01-13 Thread Matt Mahoney
I think what Mike is saying is that I could draw what I think a flying house 
would look like, and you could look at my picture and say it was a flying 
house, even though neither of us has ever seen one. Therefore, AGI should be 
able to solve the same kind of problems, and why aren't we designing and 
testing AGI this way? But don't worry about it. Mike doesn't know how to solve 
the problem either.

-- Matt Mahoney, matmaho...@yahoo.com


--- On Tue, 1/13/09, Jim Bromer  wrote:

> From: Jim Bromer 
> Subject: Re: [agi] [WAS The Smushaby] The Logic of Creativity
> To: agi@v2.listbox.com
> Date: Tuesday, January 13, 2009, 3:02 PM
> I am reluctant to say this, but I am not sure if I actually
> understand
> what Mike is getting at. He described a number of logical
> (in the
> greater sense of being reasonable and structured) methods
> by which one
> could achieve some procedural goal, and then he declares
> that logic
> (in this greater sense that I believe acknowledged) was
> incapable of
> achieving it.
> 
> Let's take a flying house.  I have to say that there
> was a very great
> chance that I misunderstood what Mike was saying, since I
> believe that
> he effectively said that a computer program, using
> logically derived
> systems could not come to the point where it could
> creatively draw a
> picture of a flying house like a child might.
> 
> If that was what he was saying then it is very strange. 
> Obviously,
> one could program a computer to draw a flying house.  So
> right away,
> his point must have been under stated, because that means
> that a
> computer program using computer logic (somewhere within
> this greater
> sense of the term) could follow a program designed to get
> it to draw a
> flying house.
> 
> So right away, Mike's challenge can't be taken
> seriously.  If we can
> use logical design to get the computer program to draw a
> flying house,
> we can find more creative ways to get it to the same point.
>  Do you
> understand what I am saying?  You aren't actually going
> to challenge
> me to write a rather insipid program that will draw a
> flying house for
> you are you?  You accept the statement that I could do that
> if I
> wanted to right?  If you do accept that statement, then you
> should be
> able to accept the fact that I could also write a more
> elaborate
> computer program to do the same thing, only it might, for
> example, do
> so only after the words "house" and
> "flying" were input. I think you
> understand that I could write a slightly more elaborate
> computer
> program to do the something like that.  Ok, now I could
> keep making it
> more complicated and eventually I could get to the point
> where where
> it could take parts of pictures that it was exposed to and
> draw them
> in more creative combinations.   If it was exposed to
> pictures of
> airplanes flying, and if it was exposed to pictures of
> houses, it
> might,. through quasi random experimentation try drawing a
> picture of
> the airplane flying past the house as if the house was an
> immense
> mountain, and then it might try some clouds as landscaping
> for the
> house and then it might try a cloud with a driveway,
> garbage can and a
> chimney, and eventually it might even draw a picture of a
> house with
> wings.  All I need to do that is to use some shape
> detecting
> algorithms that have been developed for graphics programs
> and are used
> all the time by graphic artists that can approximately
> determine the
> shape of the house and airplane in the different pictures
> and then it
> would just be a matter of time before it could (and would)
> try to draw
> a flying house.
> 
> Which step do you doubt, or did I completely misunderstand
> you?
> 1. I could (I hope I don't have to) write a program
> that could draw a
> flying house.
> 2. I could make it slightly more elaborate so, for example,
> that it
> would only draw the flying house if the words
> 'house' and 'flying'
> were input.
> 3. I could vary the program in many other ways.  Now
> suppose that I
> showed you one of these programs.  After that I could make
> it more
> complicated so that it went through a slightly more
> creative process
> than the program you saw the previous time.
> 4. I could continue to make the program more and more
> complicated. I
> could, (with a lot of graphics techniques that I know about
> but
> haven't actually mastered) write the program so that if
> it was exposed
> to pictures of houses and to pictures of flying, would have
> the
> ability to eventually draw 

Re: [agi] What Must a World Be That a Humanlike Intelligence May Develop In It?

2009-01-13 Thread Matt Mahoney
--- On Tue, 1/13/09, Ben Goertzel  wrote:

> The complexity of a simulated environment is tricky to estimate, if
> the environment contains complex self-organizing dynamics, random
> number generation, and complex human interactions ...

In fact it's not computable. But if you write 10^6 bits of code for your 
simulator, you know it's less than 10^6 bits.

But I wonder which is a better test of AI.

http://cs.fit.edu/~mmahoney/compression/text.html
is based on natural language prediction, equivalent to the Turing test. The 
data has 10^9 bits of complexity, just enough to train a human adult language 
model.

http://cs.fit.edu/~mmahoney/compression/uiq/
is based on Legg and Hutter's universal intelligence. It probably has a few 
hundred bits of complexity, designed to be just beyond the reach of 
tractability for universal algorithms like AIXI^tl.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] What Must a World Be That a Humanlike Intelligence May Develop In It?

2009-01-13 Thread Matt Mahoney
My response to Ben's paper is to be cautious about drawing conclusions from 
simulated environments. Human level AGI has an algorithmic complexity of 10^9 
bits (as estimated by Landauer). It is not possible to learn this much 
information from an environment that is less complex. If a baby AI did perform 
well in a simplified simulation of the world, it would not imply that the same 
system would work in the real world. It would be like training a language model 
on a simple, artificial language and then concluding that the system could be 
scaled up to learn English.

This is a lesson from my dissertation work in network intrusion anomaly 
detection. This was a machine learning task in which the system was trained on 
attack-free network traffic, and then identified anything out of the ordinary 
as malicious. For development and testing, we used the 1999 MIT-DARPA Lincoln 
Labs data set consisting of 5 weeks of synthetic network traffic with hundreds 
of labeled attacks. The test set developers took great care to make the data as 
realistic as possible. They collected statistics from real networks, built an 
isolated network of 4 real computers running different operating systems, and 
thousands of simulated computers that generated HTTP requests to public 
websites and mailing lists, and generated synthetic email using English word 
bigram frequencies, and other kinds of traffic.

In my work I discovered a simple algorithm that beat the best intrusion 
detection systems available at the time. I parsed network packets into 
individual 1-4 byte fields, recorded all the values that ever occurred at least 
once in training, and flagged any new value in the test data as suspicious, 
with a score inversely proportional to the size of the set of values observed 
in training and proportional to the time since the previous anomaly.

Not surprisingly, the simple algorithm failed on real network traffic. There 
were too many false alarms for it to be even remotely useful. The reason it 
worked on the synthetic traffic was that it was algorithmically simple compared 
to real traffic. For example, one of the most effective tests was the TTL 
value, a counter that decrements with each IP routing hop, intended to prevent 
routing loops. It turned out that most of the attacks were simulated from a 
machine that was one hop further away than the machines simulating normal 
traffic.

A problem like that could have been fixed, but there were a dozen others that I 
found, and probably many that I didn't find. It's not that the test set 
developers weren't careful. They spent probably $1 million developing it 
(several people over 2 years). It's that you can't simulate the high complexity 
of thousands of computers and human users with anything less than that. Simple 
problems have simple solutions, but that's not AGI.

-- Matt Mahoney, matmaho...@yahoo.com


--- On Fri, 1/9/09, Ben Goertzel  wrote:

> From: Ben Goertzel 
> Subject: [agi] What Must a World Be That a Humanlike Intelligence May Develop 
> In It?
> To: agi@v2.listbox.com
> Date: Friday, January 9, 2009, 5:58 PM
> Hi all,
> 
> I intend to submit the following paper to JAGI shortly, but
> I figured
> I'd run it past you folks on this list first, and
> incorporate any
> useful feedback into the draft I submit
> 
> This is an attempt to articulate a virtual world
> infrastructure that
> will be adequate for the development of human-level AGI
> 
> http://www.goertzel.org/papers/BlocksNBeadsWorld.pdf
> 
> Most of the paper is taken up by conceptual and
> requirements issues,
> but at the end specific world-design proposals are made.
> 
> This complements my earlier paper on AGI Preschool.  It
> attempts to
> define what kind of underlying virtual world infrastructure
> an
> effective AGI preschool would minimally require.
> 
> thx
> Ben G
> 
> 
> 
> -- 
> Ben Goertzel, PhD
> CEO, Novamente LLC and Biomind LLC
> Director of Research, SIAI
> b...@goertzel.org
> 
> "I intend to live forever, or die trying."
> -- Groucho Marx



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] fuzzy-probabilistic logic again

2009-01-12 Thread Matt Mahoney
--- On Mon, 1/12/09, YKY (Yan King Yin)  wrote:

> I have refined my P(Z) logic a bit.  Now the truth values are all
> unified to one type, probability distribution over Z, which has a
> pretty nice interpretation.  The new stuff are at sections 4.4.2 and
> 4.4.3.
> 
> http://www.geocities.com/genericai/P-Z-logic-excerpt-12-Jan-2009.pdf

Do you have any experimental results supporting your proposed probabilistic 
fuzzy logic implementation? How would you devise such an experiment (for 
example, a prediction task) to test alternative interpretations of logical 
operators like AND, OR, NOT, IF-THEN, etc? Maybe you could manually encoding 
knowledge in your system (like you did with Goldilocks) and test whether it can 
make inferences? I'd be more interested to see results on real data, however.

(Also, instead of a disclaimer about political correctness, couldn't you just 
find examples that don't reveal your obsession with sex?)

-- Matt Mahoney, matmaho...@yahoo.com




> 
> I'm wondering if anyone is interested in helping me
> implement the
> logic or develop an AGI basing on it?  I have already
> written part of
> the inference engine in Lisp.
> 
> Also, is anyone here working on fuzzy or probabilistic
> logics, other
> than Ben and Pei and me?
> 
> YKY
> 
> 
> ---
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] initial reaction to A2I2's call center product

2009-01-12 Thread Matt Mahoney
--- On Mon, 1/12/09, Ben Goertzel  wrote:

> AGI company A2I2 has released a product for automating call
> center functionality, see...
> 
> http://www.smartaction.com/index.html

It would be nice to see some transcripts of actual conversation between the 
system and customers to get some idea of how well the system actually works.

You'll notice that the contact for more information is a live human...


-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=126863270-d7b0b0
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-09 Thread Matt Mahoney
--- On Thu, 1/8/09, Vladimir Nesov  wrote:

> > I claim that K(P) > K(Q) because any description of P must include
> > a description of Q plus a description of what P does for at least one other 
> > input.
> >
> 
> Even if you somehow must represent P as concatenation of Q and
> something else (you don't need to), it's not true that always
> K(P)>K(Q). It's only true that length(P)>length(Q), and longer strings
> can easily have smaller programs that output them. If P is
> 10^(10^10) symbols X, and Q is some random number of X smaller
> than 10^(10^10), it's probably K(P) substring of P.

Well, it is true that you can find |P| < |Q| for some cases of P nontrivially 
simulating Q depending on the choice of language. However, it is not true on 
average. It is also not possible for P to nontrivially simulate itself because 
it is a contradiction to say that P does everything that Q does and at least 
one thing that Q doesn't do if P = Q.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-09 Thread Matt Mahoney
Mike, after a sequence of free associations, you drift from the original 
domain. How is that incompatible with the model I described? I use A, B, C, as 
variables to represent arbitrary thoughts.

-- Matt Mahoney, matmaho...@yahoo.com

--- On Fri, 1/9/09, Mike Tintner  wrote:
From: Mike Tintner 
Subject: Re: [agi] The Smushaby of Flatway.
To: agi@v2.listbox.com
Date: Friday, January 9, 2009, 10:08 AM


I
 
 
 _filtered #yiv455060292 {
font-family:Courier;}
 _filtered #yiv455060292 {
font-family:Tms Rmn;}
 _filtered #yiv455060292 {margin:1.0in 77.95pt 1.0in 77.95pt;}
#yiv455060292 P.MsoNormal {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 LI.MsoNormal {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 DIV.MsoNormal {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H1 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:12pt 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H2 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:6pt 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H3 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H4 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H5 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 H6 {
FONT-WEIGHT:normal;FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 P.MsoHeading7 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 LI.MsoHeading7 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 DIV.MsoHeading7 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 P.MsoHeading8 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 LI.MsoHeading8 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 DIV.MsoHeading8 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 P.MsoHeading9 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 LI.MsoHeading9 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 DIV.MsoHeading9 {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt;FONT-FAMILY:Courier;}
#yiv455060292 P.MsoNormalIndent {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt 0.5in;FONT-FAMILY:Courier;}
#yiv455060292 LI.MsoNormalIndent {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt 0.5in;FONT-FAMILY:Courier;}
#yiv455060292 DIV.MsoNormalIndent {
FONT-SIZE:12pt;MARGIN:0in 0in 0pt 0.5in;FONT-FAMILY:Courier;}
#yiv455060292 A:link {
COLOR:blue;TEXT-DECORATION:underline;}
#yiv455060292 SPAN.MsoHyperlink {
COLOR:blue;TEXT-DECORATION:underline;}
#yiv455060292 A:visited {
COLOR:purple;TEXT-DECORATION:underline;}
#yiv455060292 SPAN.MsoHyperlinkFollowed {
COLOR:purple;TEXT-DECORATION:underline;}
#yiv455060292 P.MsoPlainText {
FONT-SIZE:10pt;MARGIN:0in 0in 0pt;FONT-FAMILY:"Courier New";}
#yiv455060292 LI.MsoPlainText {
FONT-SIZE:10pt;MARGIN:0in 0in 0pt;FONT-FAMILY:"Courier New";}
#yiv455060292 DIV.MsoPlainText {
FONT-SIZE:10pt;MARGIN:0in 0in 0pt;FONT-FAMILY:"Courier New";}
#yiv455060292 DIV.Section1 {
}

Matt,
 
I mainly want to lay down a marker here for a 
future discussion.
 
What you have done  is what all AGI-ers/AI-ers 
do. Faced with the problem of domain-switching - (I pointed out that the human 
brain and human thought are * freely domain-switching*), -  you 
have simply ignored it - and, I imagine, are completely unaware that you have 
done so. And this, remember, is *the* problem of AGI - what should be 
the central focus of all discussion here.
 
If you look at your examples, you will find that 
they are all *intra-domain* and do not address domain-switching at all 
-
 
a. "if you learned the associations A-B and B-C, then A will predict C. 

> That is called "reasoning"
 
b) a word-word 
matrix M from a large text corpus, ..gives you something similar to 
> 
your free association chain like rain-wet-water-...
 
No domain-switching there.
 
Compare these 
with my 
 
b) 
domain-switching chain -"COW" - DOG - TAIL - CURRENT CRISIS - LOCAL 
VS
>> GLOBAL
>> THINKING - WHAT A NICE DAY - MUST GET ON- CANT 
SPEND MUCH
>> MORE TIME ON
>> 
THIS"
 
 (switching between the domains of - Animals - Politics/Economics - 
Weather - Personal Timetable)
 
a) your 
(extremely limited) idea of (logical) reasoning is also entirely 
intra-domain - the domain of the Alphabet,   
(A-B-C). 
 
But my creative 
and similar creative chains are analogous  to switching from say 
an Alphabet domain (A-B-C) to a Foreign Languages domain (alpha - omega)  
to a Semiotics one (symbol - sign - representation) to a Fonts one (Courier - 
Times Roman) etc. etc.  - i.e. we could all easily and spontaneously form 
such a domain-switching chain.
 
Your programs 
and all the programs ever written are still incapable of doing this - switching 
domains. This, it bears repeating, is the problem of 
AGI. 
 
Because you're 
ignor

Re: [agi] The Smushaby of Flatway.

2009-01-08 Thread Matt Mahoney
--- On Thu, 1/8/09, Vladimir Nesov  wrote:

> On Fri, Jan 9, 2009 at 12:19 AM, Matt Mahoney
>  wrote:
> > Mike,
> >
> > Your own thought processes only seem mysterious
> because you can't predict what you will think without
> actually thinking it. It's not just a property of the
> human brain, but of all Turing machines. No program can
> non-trivially model itself. (By model, I mean that P models
> Q if for any input x, P can compute the output Q(x). By
> non-trivial, I mean that P does something else besides just
> model Q. (Every program trivially models itself). The proof
> is that for P to non-trivially model Q requires K(P) >
> K(Q), where K is Kolmogorov complexity, because P needs a
> description of Q plus whatever else it does to make it
> non-trivial. It is obviously not possible for K(P) >
> K(P)).
> >
> 
> Matt, please stop. I even constructed an explicit
> counterexample to
> this pseudomathematical assertion of yours once. You
> don't pay enough
> attention to formal definitions: what this "has a
> description" means,
> and which reference TMs specific Kolmogorov complexities
> are measured
> in.

Your earlier counterexample was a trivial simulation. It simulated itself but 
did nothing else. If P did something that Q didn't, then Q would not be 
simulating P.

This applies regardless of your choice of universal TM.

I suppose I need to be more precise. I say "P simulates Q" if for all x, 
P("what is Q(x)?") = "Q(x)=y" iff Q(x)=y (where x and y are arbitrary strings). 
When I say that P does something else, I mean that it accepts at least one 
input not of the form "what is Q(x)?". I claim that K(P) > K(Q) because any 
description of P must include a description of Q plus a description of what P 
does for at least one other input.


-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-08 Thread Matt Mahoney
Mike,

Your own thought processes only seem mysterious because you can't predict what 
you will think without actually thinking it. It's not just a property of the 
human brain, but of all Turing machines. No program can non-trivially model 
itself. (By model, I mean that P models Q if for any input x, P can compute the 
output Q(x). By non-trivial, I mean that P does something else besides just 
model Q. (Every program trivially models itself). The proof is that for P to 
non-trivially model Q requires K(P) > K(Q), where K is Kolmogorov complexity, 
because P needs a description of Q plus whatever else it does to make it 
non-trivial. It is obviously not possible for K(P) > K(P)).

So if you learned the associations A-B and B-C, then A will predict C. That is 
called "reasoning".

Also, each concept is associated with thousands of other concepts, not just 
A-B. If you pick the strongest associated concept not previously activated, you 
get the semi-random thought chain you describe. You can demonstrate this with a 
word-word matrix M from a large text corpus, where M[i,j] is the degree to 
which the i'th word in the vocabulary is associated with the j'th word, as 
measured by the probability of finding both words near each other in the 
corpus. Thus, M[rain,wet] and M[wet,water] have high values because the words 
often appear in the same paragraph. Traversing related words in M gives you 
something similar to your free association chain like rain-wet-water-...

-- Matt Mahoney, matmaho...@yahoo.com


--- On Thu, 1/8/09, Mike Tintner  wrote:

> From: Mike Tintner 
> Subject: Re: [agi] The Smushaby of Flatway.
> To: agi@v2.listbox.com
> Date: Thursday, January 8, 2009, 3:54 PM
> Matt:Free association is the basic way of recalling
> memories. If you experience A followed by B, then the next
> time you experience A you will think of (or predict) B.
> Pavlov demonstrated this type of learning in animals in
> 1927.
> 
> Matt,
> 
> You're not thinking your argument through. Look
> carefully at my spontaneous
> 
> "COW" - DOG - TAIL - CURRENT CRISIS - LOCAL VS
> GLOBAL
> THINKING - WHAT A NICE DAY - MUST GET ON- CANT SPEND MUCH
> MORE TIME ON
> THIS" etc. etc"
> 
> that's not A-B association.
> 
> That's 1. A-B-C  then  2. Gamma-Delta then  3.
> Languages  then  4. Number of Lines in Letters.
> 
> IOW the brain is typically not only freely associating
> *ideas* but switching freely across, and connecting, 
> radically different *domains* in any given chain of
> association. [e.g above from Animals to Economics/Politics
> to Weather to Personal Timetable]
> 
> It can do this partly because
> 
> a) single ideas have multiple, often massively mutiple, 
> idea/domain connections in the human brain, and allow one to
> go off in any of multiple tangents/directions
> b) humans have many things - and therefore multiple domains
> - on their mind at the same time concurrently  - and can
> switch as above from the immediate subject to  some other
> pressing subject  domain (e.g. from economics/politics
> (local vs global) to the weather (what a nice day).
> 
> If your "A-B, everything-is-memory-recall" thesis
> were true, our chains-of-thought-association would be
> largely repetitive, and the domain switches inevitable..
> 
> In fact, our chains (or networks) of free association and
> domain-switching are highly creative, and each one is
> typically, from a purely technical POV, novel and
> surprising. (I have never connected TAIL and CURRENT CRISIS
> before - though Animals and Politics yes. Nor have I
> connected LOCAL VS GLOBAL THINKING before with WHAT A NICE
> DAY and the weather).
> 
> IOW I'm suggesting, the natural mode of human thought -
> and our continuous streams of association - are creative.
> And achieving such creativity is the principal problem/goal
> of AGI.
> 
> So maybe it's worth taking 20 secs. of time - producing
> your own chain-of-free-association starting say with
> "MAHONEY"  and going on for another 10 or so items
> -  and trying to figure out how the result could.possibly be
> the  narrow kind of memory-recall you're arguing for.
> It's an awful lot to ask for, but could you possibly try
> it, analyse it and report back?
> 
> [Ben claims to have heard every type of argument I make
> before,  (somewhat like your A-B memory claim), so perhaps
> he can tell me where he's read before about the Freely
> Associative, Freely Domain Switching nature of human thought
> - I'd be interested to follow up on it]. 



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-08 Thread Matt Mahoney
--- On Thu, 1/8/09, Mike Tintner  wrote:

> Matt,
> 
> Thanks. But how do you see these:
> 
> "Pattern recognition in parallel, and hierarchical
> learning of increasingly complex patterns by classical
> conditioning (association), clustering in context space
> (feature creation), and reinforcement learning to meet
> evolved goals."
> 
> as fundamentally different from logicomathematical
> thinking? ("Reinforcement learning" strikes me as
> literally extraneous and not a mode of thinking). Perhaps
> you need to explain why conditioned association is
> different.

Free association is the basic way of recalling memories. If you experience A 
followed by B, then the next time you experience A you will think of (or 
predict) B. Pavlov demonstrated this type of learning in animals in 1927. Hebb 
proposed a neural model in 1949 which has since been widely accepted. The model 
is unrelated to first order logic. It is a strengthening of the connections 
from neuron A to neuron B.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-08 Thread Matt Mahoney
--- On Thu, 1/8/09, Mike Tintner  wrote:

> What then do you see as the way people *do* think? You
> surprise me, Matt, because both the details of your answer
> here and your thinking generally strike me as *very*
> logicomathematical - with lots of emphasis on numbers and
> compression - yet you seem to be acknowledging here, like
> Jim,  the fundamental deficiencies of the logicomathematical
> form - and it is indeed only one form - of thinking. 

Pattern recognition in parallel, and hierarchical learning of increasingly 
complex patterns by classical conditioning (association), clustering in context 
space (feature creation), and reinforcement learning to meet evolved goals.

You can't write a first order logic expression that inputs a picture and tells 
you whether it is a cat or a dog. Yet any child can do it. Logic is great for 
abstract mathematics. We regard it as the highest form of thought, the hardest 
thing that humans can learn, yet it is the easiest problem to solve on a 
computer.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-08 Thread Matt Mahoney
--- On Wed, 1/7/09, Ben Goertzel  wrote:
>if proving Fermat's Last theorem was just a matter of doing math, it would 
>have been done 150 years ago ;-p
>
>obviously, all hard problems that can be solved have already been solved...
>
>???

In theory, FLT could be solved by brute force enumeration of proofs until a 
match to Wiles' is found. In theory, AGI could be solved by coding all the 
knowledge in LISP. The difference is that 50 years ago people actually expected 
the latter to work. Some people still believe so.

AGI is an engineering and policy problem. We already have small scale neural 
models of learning, language, vision, and motor control. We currently lack the 
computing power (10^16 OPS, 10^15 bits) to implement these at human levels, but 
Moore's law will take care of that.

But that is not the hard part of the problem. AGI is a system that eliminates 
our need to work, to think, and to function in the real world. Its value is USD 
10^15, the value of the global economy. Once we have the hardware, we still 
need to extract 10^18 bits of knowledge from human brains. That is the 
complexity of the global economy (assuming 10^10 people x 10^9 bits per person 
x 0.1 fraction consisting of unique job skills). This is far bigger than the 
internet. The only way to extract this knowledge without new technology like 
brain scanning is by communication at the rate of 2 bits per second per person. 
The cheapest option is a system of pervasive surveillance where everything you 
say and do is public knowledge.

AGI is too expensive for any person or group to build or own. It is a vastly 
improved internet, a communication system so efficient that the world's 
population starts to look like a single entity, and nobody notices or cares as 
silicon gradually replaces carbon.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] The Smushaby of Flatway.

2009-01-07 Thread Matt Mahoney
Logic has not solved AGI because logic is a poor model of the way people think.

Neural networks have not solved AGI because you would need about 10^15 bits of 
memory and 10^16 OPS to simulate a human brain sized network.

Genetic algorithms have not solved AGI because the computational requirements 
are even worse. You would need 10^36 bits just to model all the world's DNA, 
and even if you could simulate it in real time, it took 3 billion years to 
produce human intelligence the first time.

Probabilistic reasoning addresses only one of the many flaws of first order 
logic as a model of AGI. Reasoning under uncertainty is fine, but you haven't 
solved learning by induction, reinforcement learning, complex pattern 
recognition (e.g. vision), and language. If it was just a matter of writing the 
code, then it would have been done 50 years ago.

-- Matt Mahoney, matmaho...@yahoo.com


--- On Wed, 1/7/09, Jim Bromer  wrote:

> From: Jim Bromer 
> Subject: [agi] The Smushaby of Flatway.
> To: agi@v2.listbox.com
> Date: Wednesday, January 7, 2009, 8:23 PM
> All of the major AI paradigms, including those that are
> capable of
> learning, are flat according to my definition.  What makes
> them flat
> is that the method of decision making is
> minimally-structured and they
> funnel all reasoning through a single narrowly focused
> process that
> smushes different inputs to produce output that can appear
> reasonable
> in some cases but is really flat and lacks any structure
> for complex
> reasoning.
> 
> The classic example is of course logic.  Every proposition
> can be
> described as being either True or False and any collection
> of
> propositions can be used in the derivation of a conclusion
> regardless
> of whether the input propositions had any significant
> relational
> structure that would actually have made it reasonable to
> draw the
> definitive conclusion that was drawn from them.
> 
> But logic didn't do the trick, so along came neural
> networks and
> although the decision making is superficially distributed
> and can be
> thought of as being comprised of a structure of layer-like
> stages in
> some variations, the methodology of the system is really
> just as flat.
>  Again anything can be dumped into the neural network and a
> single
> decision making process works on the input through a
> minimally-structured reasoning system and output is
> produced
> regardless of the lack of appropriate relative structure in
> it.  In
> fact, this lack of discernment was seen as a major
> breakthrough!
> Surprise, neural networks did not work just like the mind
> works in
> spite of the years and years of hype-work that went into
> repeating
> this slogan in the 1980's.
> 
> Then came Genetic Algorithms and finally we had a system
> that could
> truly learn to improve on its previous learning and how did
> it do
> this?  It used another flat reasoning method whereby
> combinations of
> data components were processed according to one simple
> untiring method
> that was used over and over again regardless of any
> potential to see
> input as being structured in more ways than one.  Is anyone
> else
> starting to discern a pattern here?
> 
> Finally we reach the next century to find that the future
> of AI has
> already arrived and that future is probabilistic reasoning!
>  And how
> is probabilistic reasoning different?  Well, it can solve
> problems
> that logic, neural networks, genetic algorithms
> couldn't!  And how
> does probabilistic reasoning do this?  It uses a funnel
> minimally-structured method of reasoning whereby any input
> can be
> smushed together with other disparate input to produce a
> conclusion
> which is only limited by the human beings who strive to
> program it!
> 
> The very allure of minimally-structured reasoning is that
> it works
> even in some cases where it shouldn't.  It's the
> hip hooray and bally
> hoo of the smushababies of Flatway.
> 
> Jim Bromer



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-30 Thread Matt Mahoney
John,
So if consciousness is important for compression, then I suggest you write two 
compression programs, one conscious and one not, and see which one compresses 
better. 

Otherwise, this is nonsense.

-- Matt Mahoney, matmaho...@yahoo.com

--- On Tue, 12/30/08, John G. Rose  wrote:
From: John G. Rose 
Subject: RE: [agi] Universal intelligence test benchmark
To: agi@v2.listbox.com
Date: Tuesday, December 30, 2008, 9:46 AM




 
 






If the agents were p-zombies or just not conscious they would have
different motivations. 

   

Consciousness has properties of communication protocol and effects
inter-agent communication. The idea being it enhances agents' existence and
survival. I assume it facilitates collective intelligence, generally. For a
multi-agent system with a goal of compression or prediction the agent
consciousness would have to be catered.  So introducing -  

Consciousness of X is: the idea or feeling that X is
correlated with "Consciousness of X"

to
the agents would give them more "glue" if they expended that consciousness
on one another. The communications dynamics of the system would change 
verses
a similar non-conscious multi-agent system. 

   

John 

   







From: Ben Goertzel
[mailto:b...@goertzel.org] 

Sent: Monday, December 29, 2008 2:30 PM

To: agi@v2.listbox.com

Subject: Re: [agi] Universal intelligence test benchmark 





   



Consciousness of X is: the idea or feeling that X is correlated with
"Consciousness of X"



;-)



ben g 



On Mon, Dec 29, 2008 at 4:23 PM, Matt Mahoney  wrote: 



--- On Mon, 12/29/08, John G.
Rose 
wrote: 





> > What does
consciousness have to do with the rest of your argument?

> >

>

> Multi-agent systems should need individual consciousness to

> achieve advanced

> levels of collective intelligence. So if you are

> programming a multi-agent

> system, potentially a compressor, having consciousness in

> the agents could

> have an intelligence amplifying effect instead of having

> non-conscious

> agents. Or some sort of primitive consciousness component

> since higher level

> consciousness has not really been programmed yet.

>

> Agree? 



No. What do you mean by "consciousness"?



Some people use "consciousness" and intelligence"
interchangeably. If that is the case, then you are just using a circular
argument. If not, then what is the difference? 





-- Matt Mahoney, matmaho...@yahoo.com







 













  

  
  agi | Archives

 | Modify
 Your Subscription


  

  


 




---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] Universal intelligence test benchmark

2008-12-29 Thread Matt Mahoney
--- On Mon, 12/29/08, Philip Hunt  wrote:
> Am I right in understanding that the coder from fpaq0 could
> be used with any other predictor?

Yes. It has a simple interface. You have a class called Predictor which is your 
bit sequence predictor. It has 2 member functions that you have to write. p() 
should return your estimated probability that the next bit will be a 1, as a 12 
bit number (0 to 4095). update(y) then tells you what that bit actually was, a 
0 or 1. The encoder will alternately call these 2 functions for each bit of the 
sequence. The predictor doesn't know whether it is compressing or decompressing 
because it sees exactly the same sequence either way.

So the easy part is done :)

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] [Science Daily] Our Unconscious Brain Makes The Best Decisions Possible

2008-12-29 Thread Matt Mahoney
--- On Mon, 12/29/08, Richard Loosemore  wrote:

> 8-) Don't say that too loudly, Yudkowsky might hear
> you. :-)
...
> When I suggested that someone go check some of his ravings
> with an outside authority, he banned me from his discussion
> list.

Yudkowsky's side of the story might be of interest...

http://www.sl4.org/archive/0608/15895.html
http://www.sl4.org/archive/0608/15928.html

-- Matt Mahoney, matmaho...@yahoo.com


> From: Richard Loosemore 
> Subject: Re: [agi] [Science Daily] Our Unconscious Brain Makes The Best 
> Decisions Possible
> To: agi@v2.listbox.com
> Date: Monday, December 29, 2008, 4:02 PM
> Lukasz Stafiniak wrote:
> >
> http://www.sciencedaily.com/releases/2008/12/081224215542.htm
> > 
> > Nothing surprising ;-)
> 
> Nothing surprising?!!
> 
> 8-) Don't say that too loudly, Yudkowsky might hear
> you. :-)
> 
> The article is a bit naughty when it says, of Tversky and
> Kahnemann, that "...this has become conventional wisdom
> among cognition researchers."  Actually, the original
> facts were interpreted in a variety of ways, some of which
> strongly disagreed with T & K's original
> intepretation, just like this one you reference above.  The
> only thing that is conventional wisdom is that the topic
> exists, and is the subject of dispute.
> 
> And, as many people know, I made the mistake of challenging
> Yudkowsky on precisely this subject back in 2006, when he
> wrote an essay strongly advocating T&K's original
> intepretation.  Yudkowsky went completely berserk, accused
> me of being an idiot, having no brain, not reading any of
> the literature, never answering questions, and generally
> being something unspeakably worse than a slime-oozing crank.
>  He literally wrote an essay denouncing me as equivalent to
> a flat-earth believing crackpot.
> 
> When I suggested that someone go check some of his ravings
> with an outside authority, he banned me from his discussion
> list.
> 
> Ah, such are the joys of being speaking truth to power(ful
> idiots).
> 
> ;-)
> 
> As far as this research goes, it sits somewhere down at the
> lower end of the available theories.  My friend Mike
> Oaksford in the UK has written several papers giving a
> higher level cognitive theory that says that people are, in
> fact, doing something like bayesian estimation when then
> make judgments.  In fact, people are very good at being
> bayesians, contra the loud protests of the I Am A Bayesian
> Rationalist crowd, who think they were the first to do it.
> 
> 
> 
> 
> 
> Richard Loosemore



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-29 Thread Matt Mahoney
--- On Mon, 12/29/08, John G. Rose  wrote:

> > What does consciousness have to do with the rest of your argument?
> > 
> 
> Multi-agent systems should need individual consciousness to
> achieve advanced
> levels of collective intelligence. So if you are
> programming a multi-agent
> system, potentially a compressor, having consciousness in
> the agents could
> have an intelligence amplifying effect instead of having
> non-conscious
> agents. Or some sort of primitive consciousness component
> since higher level
> consciousness has not really been programmed yet. 
> 
> Agree?

No. What do you mean by "consciousness"?

Some people use "consciousness" and intelligence" interchangeably. If that is 
the case, then you are just using a circular argument. If not, then what is the 
difference?

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-29 Thread Matt Mahoney
--- On Mon, 12/29/08, John G. Rose  wrote:

> Well that's a question. Does death somehow enhance a
> lifeforms' collective intelligence?

Yes, by weeding out the weak and stupid.

> Agents competing over finite resources.. I'm wondering if
> there were multi-agent evolutionary genetics going on would there be a
> finite resource of which there would be a relation to the collective goal of
> predicting the next symbol.

No, prediction is a secondary goal. The primary goal is to have a lot of 
descendants.

> Agent knowledge is not only passed on in their
> genes, it is also passed around to other agents Does agent death hinder
> advances in intelligence or enhance it? And then would the intelligence
> collected thus be applicable to the goal. And if so, consciousness may be
> valuable.

What does consciousness have to do with the rest of your argument?

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] Universal intelligence test benchmark

2008-12-29 Thread Matt Mahoney
--- On Mon, 12/29/08, Philip Hunt  wrote:

> Incidently, reading Matt's posts got me interested in writing a
> compression program using Markov-chain prediction. The prediction bit
> was a piece of piss to write; the compression code is proving
> considerably more difficult.

Well, there is plenty of open source software.
http://cs.fit.edu/~mmahoney/compression/

If you want to write your own model and just need a simple arithmetic coder, 
you probably want fpaq0. Most of the other programs on this page use the same 
coder or some minor variation of it.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] Universal intelligence test benchmark

2008-12-29 Thread Matt Mahoney
--- On Sun, 12/28/08, Philip Hunt  wrote:

> > Please remember that I am not proposing compression as
> > a solution to the AGI problem. I am proposing it as a
> > measure of progress in an important component (prediction).
> 
> Then why not cut out the middleman and measure prediction
> directly?

Because a compressor proves the correctness of the measurement software at no 
additional cost in either space or time complexity or software complexity. The 
hard part of compression is modeling. Arithmetic coding is essentially a solved 
problem. A decompressor uses exactly the same model as a compressor. In high 
end compressors like PAQ, the arithmetic coder takes up about 1% of the 
software, 1% of the CPU time, and less than 1% of memory.

In speech recognition research it is common to use word perplexity as a measure 
of the quality of a language model. Experimentally, it correlates well with 
word error rate. Perplexity is defined as 2^H where H is the average number of 
bits needed to encode a word. Unfortunately this is sometimes done in 
nonstandard ways, such as with restricted vocabularies and different methods of 
handling words outside the vocabulary, parsing, stemming, capitalization, 
punctuation, spacing, and numbers. Without accounting for this additional data, 
it makes published results difficult to compare. Compression removes the 
possibility of such ambiguities.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] Universal intelligence test benchmark

2008-12-28 Thread Matt Mahoney
--- On Sun, 12/28/08, Philip Hunt  wrote:

> Now, consider if I build a program that can predict how
> some sequences will continue. For example, given
> 
>ABACADAEA
> 
> it'll predict the next letter is "F", or
> given:
> 
>   1 2 4 8 16 32
> 
> it'll predict the next number is 64.

Please remember that I am not proposing compression as a solution to the AGI 
problem. I am proposing it as a measure of progress in an important component 
(prediction). Neither zip nor any entry in the Loebner contest will predict the 
next item in these sequences because they aren't very intelligent. The 
challenge for you is to solve problems like this. If you write a Loebner prize 
entry that does, it has a greater chance of winning. If you write a compressor 
that does, it will compress smaller because it will be able to assign smaller 
codes to the predicted symbols.

Random Turing machines are more likely to generate sequences with recognizable 
patterns than sequences without. That is my justification for testing with such 
data. There are many machine learning algorithms that are faster than AIXI^tl 
(randomly guessing machines) at recognizing these patterns. Obviously we must 
use some of them, or we could never solve such problems. The challenge for you 
is to discover these algorithms.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-28 Thread Matt Mahoney
--- On Sun, 12/28/08, John G. Rose  wrote:

> So maybe for improved genetic
> algorithms used for obtaining max compression there needs to be a
> consciousness component in the agents? Just an idea I think there is
> potential for distributed consciousness inside of command line compressors
> :)

No, consciousness (as the term is commonly used) is the large set of properties 
of human mental processes that distinguish life from death, such as ability to 
think, learn, experience, make decisions, take actions, communicate, etc. It is 
only relevant as an independent concept to agents that have a concept of death 
and the goal of avoiding it. The only goal of a compressor is to predict the 
next input symbol. 

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: [agi] Universal intelligence test benchmark

2008-12-27 Thread Matt Mahoney
5294A52000
1 FC4A473E25239F1291C000
1 FC4A0941282504A09400
1 FC4A00

The best compressors will compress this data to just under 3 MB, which implies 
an average algorithmic complexity of less than 24 bits per string. However, the 
language allows the construction of arbitrary 128 bit strings in a fairly 
straightforward manner.

-- Matt Mahoney, matmaho...@yahoo.com





---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-27 Thread Matt Mahoney
--- On Sat, 12/27/08, John G. Rose  wrote:

> Well I think consciousness must be some sort of out of band intelligence
> that bolsters an entity in terms of survival. Intelligence probably
> stratifies or optimizes in zonal regions of similar environmental
> complexity, consciousness being one or an overriding out-of-band one...

No, consciousness only seems mysterious because human brains are programmed 
that way. For example, I should logically be able to convince you that "pain" 
is just a signal that reduces the probability of you repeating whatever actions 
immediately preceded it. I can't do that because emotionally you are convinced 
that "pain is real". Emotions can't be learned the way logical facts can, so 
emotions always win. If you could accept the logical consequences of your brain 
being just a computer, then you would not pass on your DNA. That's why you 
can't.

BTW the best I can do is believe both that consciousness exists and 
consciousness does not exist. I realize these positions are inconsistent, and I 
leave it at that.

> > I was hoping to discover an elegant theory for AI. It didn't quite work
> > that way. It seems to be a kind of genetic algorithm: make random
> > changes to the code and keep the ones that improve compression.
> > 
> 
> Is this true for most data? For example would PI digit compression attempts
> result in genetic emergences the same as say compressing environmental
> noise? I'm just speculating that genetically originated data would require
> compression avenues of similar algorithmic complexity descriptors, for
> example PI digit data does not originate genetically so compression attempts
> would not show genetic emergences as "chained" as say environmental
> noise basically I'm asking if you can tell the difference from data that
> has a genetic origination ingredient verses all non-genetic...

No, pi can be compressed to a simple program whose size is dominated by the log 
of the number of digits you want.

For text, I suppose I should be satisfied that a genetic algorithm compresses 
it, except for the fact that so far the algorithm requires a human in the loop, 
so it doesn't solve the AI problem.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


RE: [agi] Universal intelligence test benchmark

2008-12-27 Thread Matt Mahoney
--- On Sat, 12/27/08, John G. Rose  wrote:

> > > How does consciousness fit into your compression
> > > intelligence modeling?
> > 
> > It doesn't. Why is consciousness important?
> > 
> 
> I was just prodding you on this. Many people on this list talk about the
> requirements of consciousness for AGI and I was imagining some sort of
> consciousness in one of your command line compressors :) 
> I've yet to grasp
> the relationship between intelligence and consciousness though lately I
> think consciousness may be more of an evolutionary social thing. Home grown
> digital intelligence, since it is a loner, may not require "much"
> consciousness IMO..

What we commonly call consciousness is a large collection of features that 
distinguish living human brains from dead human brains: ability to think, 
communicate, perceive, make decisions, learn, move, talk, see, etc. We only 
attach significance to it because we evolved, like all animals, to fear a large 
set of things that can kill us.

> > > Max compression implies hacks, kludges and a
> large decompressor.
> > 
> > As I discovered with the large text benchmark.
> > 
> 
> Yep and the behavior of the metrics near max theoretical
> compression is erratic I think?

It shouldn't be. There is a well defined (but possibly not computable) limit 
for each of the well defined universal Turing machines that the benchmark 
accepts (x86, C, C++, etc).

I was hoping to discover an elegant theory for AI. It didn't quite work that 
way. It seems to be a kind of genetic algorithm: make random changes to the 
code and keep the ones that improve compression.

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


Re: Spatial indexing (was Re: [agi] Universal intelligence test benchmark)

2008-12-26 Thread Matt Mahoney
--- On Sat, 12/27/08, Matt Mahoney  wrote:

> In my thesis, I proposed a vector space model where
> messages are routed in O(n) time over n nodes.

Oops, O(log n).

-- Matt Mahoney, matmaho...@yahoo.com



---
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com


  1   2   3   4   5   6   7   8   9   10   >