Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Charles Hixson
Hector Zenil wrote: On Mon, Dec 1, 2008 at 6:20 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: On Sun, Nov 30, 2008 at 11:48 PM, Hector Zenil <[EMAIL PROTECTED]> wrote: On Mon, Dec 1, 2008 at 4:55 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: But I don't get your point at all, becaus

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread J. Andrew Rogers
On Nov 30, 2008, at 7:31 AM, Philip Hunt wrote: 2008/11/30 Ben Goertzel <[EMAIL PROTECTED]>: In general, the standard AI methods can't handle pattern recognition problems requiring finding complex interdependencies among multiple variables that are obscured among scads of other variables T

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 6:20 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > On Sun, Nov 30, 2008 at 11:48 PM, Hector Zenil <[EMAIL PROTECTED]> wrote: >> On Mon, Dec 1, 2008 at 4:55 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: >>> But I don't get your point at all, because the whole idea of >>> "nondet

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
On Sun, Nov 30, 2008 at 11:48 PM, Hector Zenil <[EMAIL PROTECTED]> wrote: > On Mon, Dec 1, 2008 at 4:55 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: >> But I don't get your point at all, because the whole idea of >> "nondeterministic" randomness has nothing to do with physical >> reality... > > I do

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 4:55 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > But I don't get your point at all, because the whole idea of > "nondeterministic" randomness has nothing to do with physical > reality... I don't get it. You don't think that quantum mechanics is part of our physical reality

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 4:55 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > But I don't get your point at all, because the whole idea of > "nondeterministic" randomness has nothing to do with physical > reality... It has all to do when it is about quantum mechanics. Quantum mechanics is non-determin

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
But I don't get your point at all, because the whole idea of "nondeterministic" randomness has nothing to do with physical reality... true random numbers are uncomputable entities which can never be existed, and any finite series of observations can be modeled equally well as the first N bits of an

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 4:53 AM, Hector Zenil <[EMAIL PROTECTED]> wrote: > On Mon, Dec 1, 2008 at 4:44 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: >> OTOH, there is no possible real-world test to distinguish a "true >> random" sequence from a high-algorithmic-information quasi-random >> sequence

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 4:44 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > OTOH, there is no possible real-world test to distinguish a "true > random" sequence from a high-algorithmic-information quasi-random > sequence I know, but the point is not whether we can distinguish it, but that quantu

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Hector Zenil
On Mon, Dec 1, 2008 at 3:09 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: >> But quantum theory does appear to be directly related to limits of the >> computations of physical reality. The uncertainty theory and the >> quantization of quantum states are limitations on what can be computed by >> phys

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
OTOH, there is no possible real-world test to distinguish a "true random" sequence from a high-algorithmic-information quasi-random sequence So I don't find this argument very convincing... On Sun, Nov 30, 2008 at 10:42 PM, Hector Zenil <[EMAIL PROTECTED]> wrote: > On Mon, Dec 1, 2008 at 3:09

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
HI, > "In quantum physics, the Heisenberg uncertainty principle states that the > values of certain pairs of conjugate variables (position and momentum, for > instance) cannot both be known with arbitrary precision. That is, the more > precisely one variable is known, the less precisely the other

RE: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ed Porter
Regarding the uncertainty principal, Wikipedia says: "In quantum physics, the Heisenberg uncertainty principle states that the values of certain pairs of conjugate variables (position and momentum, for instance) cannot both be known with arbitrary precision. That is, the more precisely one vari

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Trent Waddington
On Mon, Dec 1, 2008 at 11:19 AM, Ed Porter <[EMAIL PROTECTED]> wrote: > You said "QUANTUM THEORY REALLY HAS NOTHING DIRECTLY TO DO WITH > UNCOMPUTABILITY." Please don't quote people using this style, it hurts my eyes. > But quantum theory does appear to be directly related to limits of the > comp

Re: >> RE: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
> But quantum theory does appear to be directly related to limits of the > computations of physical reality. The uncertainty theory and the > quantization of quantum states are limitations on what can be computed by > physical reality. Not really. They're limitations on what measurements of phy

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Jim Bromer
Ed, I think that we must rely on large collections of relatively simple patterns that are somehow capable of being mixed and used in interactions with the others. These interacting patterns (to use your term) would have extensive variations to make them flexible and useful with other patterns. Wh

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Jim Bromer
Charles, I don't agree with the details, but I do agree that something that is effectively similar to your description does play a role. I seem to pick a few words at a time which are following some simple plan, and yes they do go through some filters. But I think I am also selecting words as wel

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Jim Bromer
I realized that my idea of declarative-like statements could refer to statistical objects and methods as well. In fact, if they were to provide the sort of efficacy I want for them, some would have to. I am not specifically talking about mixing logic with probability theory. Thanks for the comme

Re: FW: [agi] A paper that actually does solve the problem of consciousness

2008-11-30 Thread Ben Goertzel
Ed, Unfortunately to reply to your message in detail would absorb a lot of time, because there are two issues mixed up 1) you don't know much about computability theory, and educating you on it would take a lot of time (and is not best done on an email list) 2) I may not have expressed some of m

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Stephen Reed
Pei, Matt Taylor's work at Cycorp was not closely related to his published work at AGI-08. Matt contributed to a variety of other Transfer Learning tasks, and I cannot recall exactly what those were. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://tex

Re: [agi] AIXI (was: Mushed Up Decision Processes)

2008-11-30 Thread Richard Loosemore
Philip Hunt wrote: 2008/11/29 Matt Mahoney <[EMAIL PROTECTED]>: The general problem of detecting overfitting is not computable. The principle according to Occam's Razor, formalized and proven by Hutter's AIXI model, is to choose the shortest program (simplest hypothesis) that generates the data.

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Pei Wang
Stephen, Does that mean what you did at Cycorp on transfer learning is similar to what Taylor presented to AGI-08? Pei On Sun, Nov 30, 2008 at 1:01 PM, Stephen Reed <[EMAIL PROTECTED]> wrote: > Matt Taylor was also an intern at Cycorp where was on Cycorp's Transfer > Learning team with me. > -St

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Ben Goertzel
> > Regarding winning a DARPA contract, I believe that teaming with an > established contractor, e.g. SAIC, SRI, is beneficial. > > Cheers, > -Steve Yeah, I've tried that approach too ... As it happens, I've had significant more success getting funding from various other government agencies ... b

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Stephen Reed
Matt Taylor was also an intern at Cycorp where was on Cycorp's Transfer Learning team with me. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: P

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Stephen Reed
Ben, Cycorp participated in the DARPA Transfer Learning project, as a subcontractor. My project role was simply a team member and I did not attend any PI meetings. But I did work on getting a Quake III Arena environment working at Cycorp which was to be a transfer learning testbed. I also

Re: [agi] Seeking CYC critiques

2008-11-30 Thread Stephen Reed
Robin, While I was at Cycorp, a concerted effort was made to address Vaughan Pratt's test questions. I recall that most of them required the addition of facts and rules into the Cyc KB so that they would answer. I believe that a substantial portion are included in the Cyc query regression tes

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Pei Wang
On Sun, Nov 30, 2008 at 11:17 AM, Ben Goertzel <[EMAIL PROTECTED]> wrote: > There was a DARPA program on "transfer learning" a few years back ... > I believe I applied and got rejected (with perfect marks on the > technical proposal, as usual ...) ... I never checked to see who got > the $$ and wha

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Ben Goertzel
There was a DARPA program on "transfer learning" a few years back ... I believe I applied and got rejected (with perfect marks on the technical proposal, as usual ...) ... I never checked to see who got the $$ and what they did with it... ben g On Sun, Nov 30, 2008 at 11:12 AM, Philip Hunt <[EMAI

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Philip Hunt
2008/11/30 Ben Goertzel <[EMAIL PROTECTED]>: > Hi, > >> I have proposed a problem domain called "function predictor" whose >> purpose is to allow an AI to learn across problem sub-domains, >> carrying its learning from one domain to another. (See >> http://www.includipedia.com/wiki/User:Cabalamat/F

Re: [agi] AIXI (was: Mushed Up Decision Processes)

2008-11-30 Thread Ben Goertzel
AIXI is a purely theoretic construct, requiring infinite computational resources AIXItl is a version that could be implemented in principle, but not in practice due to truly insane computational resource requirements Whether the line of thinking and body of theory underlying these things can be u

[agi] AIXI (was: Mushed Up Decision Processes)

2008-11-30 Thread Philip Hunt
2008/11/29 Matt Mahoney <[EMAIL PROTECTED]>: > > The general problem of detecting overfitting is not computable. The principle > according to Occam's Razor, formalized and proven by Hutter's AIXI model, is > to choose the shortest program (simplest hypothesis) that generates the data. > Overfitt

[agi] Re: Glocal memory

2008-11-30 Thread Ben Goertzel
A little more poking around reveals further evidence that supports the glocal model of brain memory (they talk about a "distributed plus hub" model, which is part of the glocality idea, though missing the nonlinear-attractor aspect that I think is critical to distributed memory) http://brain.guide

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Ben Goertzel
Hi, > I have proposed a problem domain called "function predictor" whose > purpose is to allow an AI to learn across problem sub-domains, > carrying its learning from one domain to another. (See > http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) > > I also think it would be use

Re: [agi] Mushed Up Decision Processes

2008-11-30 Thread Philip Hunt
2008/11/30 Ben Goertzel <[EMAIL PROTECTED]>: >> Could you give me a little more detail about your thoughts on this? >> Do you think the problem of increasing uncomputableness of complicated >> complexity is the common thread found in all of the interesting, >> useful but unscalable methods of AI? >

Re: [agi] Seeking CYC critiques

2008-11-30 Thread Robin Hanson
Hi Stephen, nice to meet you.  When I search the web for critiques of CYC, I can only find stuff from '90-95.  If no one has written critiques of CYC since then, perhaps you could comment on how applicable those early critiques would be to the current system.  For example, would CYC today at lea