David, What are the different ways you are thinking of for measuring the predictiveness? I can think of a few different possibilities (such as measuring number incorrect vs measuring fraction incorrect, et cetera) but I'm wondering which variations you consider significant/troublesome/etc.
--Abram On Thu, Jul 22, 2010 at 7:12 PM, David Jones <davidher...@gmail.com> wrote: > It's certainly not as simple as you claim. First, assigning a probability > is not always possible, nor is it easy. The factors in calculating that > probability are unknown and are not the same for every instance. Since we do > not know what combination of observations we will see, we cannot have a > predefined set of probabilities, nor is it any easier to create a > probability function that generates them for us. That is just as exactly > what I meant by quantitatively define the predictiveness... it would be > proportional to the probability. > > Second, if you can define a program ina way that is always simpler when it > is smaller, then you can do the same thing without a program. I don't think > it makes any sense to do it this way. > > It is not that simple. If it was, we could solve a large portion of agi > easily. > > On Thu, Jul 22, 2010 at 3:16 PM, Matt Mahoney <matmaho...@yahoo.com> > wrote: > > David Jones wrote: > > > But, I am amazed at how difficult it is to quantitatively define more > predictive and simpler for specific problems. > > It isn't hard. To measure predictiveness, you assign a probability to each > possible outcome. If the actual outcome has probability p, you score a > penalty of log(1/p) bits. To measure simplicity, use the compressed size of > the code for your prediction algorithm. Then add the two scores together. > That's how it is done in the Calgary challenge > http://www.mailcom.com/challenge/ and in my own text compression > benchmark. > > > > -- Matt Mahoney, matmaho...@yahoo.com > > *From:* David Jones <davidher...@gmail.com> > *To:* agi <agi@v2.listbox.com> > *Sent:* Thu, July 22, 2010 3:11:46 PM > *Subject:* Re: [agi] Re: Huge Progress on the Core of AGI > > Because simpler is not better if it is less predictive. > > On Thu, Jul 22, 2010 at 1:21 PM, Abram Demski <abramdem...@gmail.com> > wrote: > > Jim, > > Why more predictive *and then* simpler? > > --Abram > > On Thu, Jul 22, 2010 at 11:49 AM, David Jones <davidher...@gmail.com> > wrote: > > An Update.... > > I think the following gets to the heart of general AI and what it takes to > achieve it. It also provides us with evidence as to why general AI is so > difficult. With this new knowledge in mind, I think I will be much more > capable now of solving the problems and making it work. > > I've come to the conclusion lately that the best hypothesis is better > because it is more predictive and then simpler than other hypotheses (in > that order.... more predictive... then simpler). But, I am amazed at how > difficult it is to quantitatively define more predictive and simpler for > specific problems. This is why I have sometimes doubted the truth of the > statement. > > In addition, the observations that the AI gets are not representative of > all observations! This means that if your measure of "predictiveness" > depends on the number of certain observations, it could make mistakes! So, > the specific observations you are aware of may be unrepresentative of the > predictiveness of a hypothesis relative to the truth. If you try to > calculate which hypothesis is more predictive and you don't have the > critical observations that would give you the right answer, you may get the > wrong answer! This all depends of course on your method of calculation, > which is quite elusive to define. > > Visual input from screenshots, for example, can be somewhat malicious. > Things can move, appear, disappear or occlude each other suddenly. So, > without sufficient knowledge it is hard to decide whether matches you find > between such large changes are because it is the same object or a different > object. This may indicate that bias and preprogrammed experience should be > introduced to the AI before training. Either that or the training inputs > should be carefully chosen to avoid malicious input and to make them nice > for learning. > > This is the "correspondence problem" that is typical of computer vision and > has never been properly solved. Such malicious input also makes it difficult > to learn automatically because the AI doesn't have sufficient experience to > know which changes or transformations are acceptable and which are not. It > is immediately bombarded with malicious inputs. > > I've also realized that if a hypothesis is more "explanatory", it may be > better. But quantitatively defining explanatory is also elusive and truly > depends on the specific problems you are applying it to because it is a > heuristic. It is not a true measure of correctness. It is not loyal to the > truth. "More explanatory" is really a heuristic that helps us find > hypothesis that are more predictive. The true measure of whether a > hypothesis is better is simply the most accurate and predictive hypothesis. > That is the ultimate and true measure of correctness. > > Also, since we can't measure every possible prediction or every last > prediction (and we certainly can't predict everything), our measure of > predictiveness can't possibly be right all the time! We have no choice but > to use a heuristic of some kind. > > So, its clear to me that the right hypothesis is "more predictive and then > simpler". But, it is also clear that there will never be a single measure of > this that can be applied to all problems. I hope to eventually find a nice > model for how to apply it to different problems though. This may be the > reason that so many people have tried and failed to develop general AI. Yes, > there is a solution. But there is no silver bullet that can be applied to > all problems. Some methods are better than others. But I think another major > reason of the failures is that people think they can predict things without > sufficient information. By approaching the problem this way, we compound the > need for heuristics and the errors they produce because we simply don't have > sufficient information to make a good decision with limited evidence. If > approached correctly, the right solution would solve many more problems with > the same efforts than a poor solution would. It would also eliminate some of > the difficulties we currently face if sufficient data is available to learn > from. > > In addition to all this theory about better hypotheses, you have to add on > the need to solve problems in reasonable time. This also compounds the > difficulty of the problem and the complexity of solutions. > > I am always fascinated by the extraordinary difficulty and complexity of > this problem. The more I learn about it, the more I appreciate it. > > Dave > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> <http://www.listbox.com> > > > > -- > Abram Demski > http://lo-tho.blogspot.com/ > http://groups.google.com/group/one-logic > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> <http://www.listbox.com> > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> <http://www.listbox.com> > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> <http://www.listbox.com> > > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Abram Demski http://lo-tho.blogspot.com/ http://groups.google.com/group/one-logic ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c Powered by Listbox: http://www.listbox.com