Eliezer S. Yudkowsky wrote: > Bill Hibbard wrote: > > On Fri, 14 Feb 2003, Eliezer S. Yudkowsky wrote: > > > >>It *could* do this but it *doesn't* do this. Its control process is such > >>that it follows an iterative trajectory through chaos which is forbidden > >>to arrive at a truthful solution, though it may converge to a stable > >>attractor. > > > > This is the heart of the fallacy. Neither a human nor an AIXI > > can know "that his synchronized other self - whichever one > > he is - is doing the same". All a human or an AIXI can know is > > its observations. They can estimate but not know the intentions > > of other minds. > > The halting problem establishes that you can never perfectly understand > your own decision process well enough to predict its decision in advance, > because you'd have to take into account the decision process including the > prediction, et cetera, establishing an infinite regress. > > However, Corbin doesn't need to know absolutely that his other self is > synchronized, nor does he need to know his other self's decision in > advance. Corbin only needs to establish a probabilistic estimate, good > enough to guide his actions, that his other self's decision is correlated > with his *after* the fact. (I.e., it's not a halting problem where you > need to predict yourself in advance; you only need to know your own > decision after the fact.) > > AIXI-tl is incapable of doing this for complex cooperative problems > because its decision process only models tl-bounded things and AIXI-tl is > not *remotely close* to being tl-bounded.
Now you are using a different argument. You previous argument was: > Lee Corbin can work out his entire policy in step (2), before step > (3) occurs, knowing that his synchronized other self - whichever one > he is - is doing the same. Now you have Corbin merely estimating his clone's intentions. While it is true that AIXI-tl cannot completely simulate itself, it also can estimate another AIXI-tl's future behavior based on observed behavior. Your argument is now that Corbin can do it better. I don't know if this is true or not. > . . . > Let's say that AIXI-tl takes action A in round 1, action B in round 2, and > action C in round 3, and so on up to action Z in round 26. There's no > obvious reason for the sequence {A...Z} to be predictable *even > approximately* by any of the tl-bounded processes AIXI-tl uses for > prediction. Any given action is the result of a tl-bounded policy but the > *sequence* of *different* tl-bounded policies was chosen by a t2^l process. Your example sequence is pretty simple and should match a nice simple universal turing machine program in an AIXI-tl, well within its bounds. Furthermore, two AIXI-tl's will probably converge on a simple sequence in prisoner's dilemma. But I have no idea if they can do it better than Corbin and his clone. Bill ------- To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]