The most effective image recognition systems we have today, including semantic understanding of scenes, are all NN variants to the best of my knowledge.
-s On 03/08/2015 08:51 PM, Jim Bromer via AGI wrote: > > Jim, can you describe an algorithm where P = NP would exponentially >> speed up visual processing? > The complexity of trying to find how each pixel (or tiny area) relates > to other pixels is not a true Satisfiability problem until a (more > classic-like) image analysis algorithm gathers some information. For > example, is some area a part of the background? That is not an easy > question because the 'background' may be quite diverse. Furthermore, > the desired 'subjects' of an image may not be -perceived- before image > analysis gets going. Once an algorithm starts picking up information > than the variations of possibilities starts forming a true > Satisfiaibilty problem although it may not be expressed in simple > Boolean relations. > > While most image analysis would take place across a field of data that > does not mean that all image analysis are essentially neural networks. > > I don't want to dismiss deep learning neural networks just because > they have not achieved even shallow AGI. But look at character > recognition. Alphabet characters have flat distinct shapes. Although > they may vary widely, one might still design a classical algorithm > that defines how near a subject character is to a set of training > characters in the terms of vectors (and weighted reasoning) or > something like that. The attempt to use of vectors with 'ideas' or > semantic objects has been inadequate because the domains of 'ideas' > do not all fit into a domain of vector space. Space and much of > physics have benefited from a great deal of effective mathematical > analysis over the centuries. So computers are great at predicting the > weather because the averages of history (of different measures of > events) can be combined with knowledge of the causes of the weather > and presto - you have some great weather predictions. But if you > wanted to push deep anns what are you going to find? You are going to > find that you have to push up against the complexity barriers. > Although the Boolean relations between areas of an image in an ann may > be hidden, they are never the less creating complexity barriers. Anns > work so well because some problems can be effectively solved by using > multiple metric approximations. > > But, let's suppose that AGI will one day be accomplished using metrics > - in other words weighted reasoning, probability and so on. I realize > that it is a real possibility. That means that image analysis > algorithms will be able to recognize anything a human being could > recognize. Here the problem is not a question of taking tens or > hundreds of flat land characteristics for tens of characters and > coming up with a method that effectively measures how close a subject > character is to different kinds of metrics derived from the training > characters and then picks the closest matches, but of taking thousands > of characteristics from thousands of objects to derive estimates of > what is pictured. In this case the problem is combinatorially more > complex just because there are so many more possible subjects and > variants of those subjects. It should be clear that this is a true > satisfiability problem even though Anns don't deal directly with the > satisfiability issue because the acquired weights are so heavily > distributed. The complexity barriers are effectively satisfiability > problems even though the distinct relations may be hidden. I hope this > makes some sense because I really don't know that much about deep > learning Anns and I haven't really done that much image analysis. > > One other thing. As you know, a neural network is not the only kind of > algorithm that can learn. So it is pretty easy to imagine an algorithm > that is based on supervised learning to develop discreet relations > that form networks. The relations would of course include weights and > so on. One of the reasons I would like to develop some better SAT > methods is that I then could develop AI models using simple concepts > and build on it. Right now the complexity barriers are so low and so > pervasive that you can't get any real traction unless the problem can > be solved using weighted approximations. So weighted reasoning is way > ahead right now but that may not always be the case. My goal is not to > achieve detailed precision but to get some basic traction by starting > with something simple and improving it. > > Jim Bromer > > > On Fri, Mar 6, 2015 at 10:04 AM, Matt Mahoney via AGI <[email protected]> > wrote: >> Jim, can you describe an algorithm where P = NP would exponentially >> speed up visual processing? My understanding is that the most advanced >> vision algorithms use deep neural networks with a structure similar to >> the visual cortex. In general, neural network size (in synapses) >> should be proportional to the training set information content. Thus, >> training time is O(n^2). >> >> On Thu, Mar 5, 2015 at 10:01 PM, Jim Bromer via AGI <[email protected]> wrote: >>> Matt said: >>> Vision is a pattern recognition problem. You input a picture of a cat >>> and output a label like "cat". It is not NP-complete because (1) >>> experimentally, the problem scales polynomially with input size and >>> (2) the time to verify that a label like "cat" is correct is about the >>> same as the time it takes to label the image. Thus, the problem is in >>> P and would not benefit even if P = NP. >>> ------------------------------------------------- >>> This is a truly insipid response. You have taken one narrow situation >>> and used it in an over-generalization of a kind of AGI problem. "The >>> problem scales polynomially with input size? The point that I made is >>> that the general analysis of imagery is presumably as bad or worse >>> than NP (in the lexicon of the day). What I mean is that there is >>> sufficient evidence that AGI is, in the worse case, at least >>> exponentially difficult and that makes it worthwhile to examine why >>> that may be. One reason, the reason I gave, is that the easiest >>> methods to make a methodical and thorough analysis of the relations >>> between associated pixels would be those that are (literally) in NP. >>> The implied case of scaling a particular picture and arguing that it >>> would scale polynomially with input size is analogous to saying that >>> converting an unrestricted Boolean Satisfiability problem to 3-SAT >>> scales polynomially (and that somehow proves that unrestricted SAT >>> scales polynomially). It is pretty obvious that you have little >>> experience with visual data. >>> >>> This is an example of a blatant overgeneralization being declared as >>> if it were a factual statement. I can't casually explain why visual >>> analysis is at least exponentially difficult because I am not enough >>> of an expert to be that familiar with all the problems. However, I am >>> confident that there is no overwhelming evidence to suggest that, in >>> general, it is less difficult. >>> Jim Bromer >>> >>> >>> On Thu, Mar 5, 2015 at 1:04 PM, Matt Mahoney via AGI <[email protected]> >>> wrote: >>>> On Wed, Mar 4, 2015 at 2:51 AM, Jim Bromer <[email protected]> wrote: >>>>> On Tue, Feb 17, 2015 at 11:52 PM, Matt Mahoney via AGI >>>>> <[email protected]> wrote: >>>>>> On Tue, Feb 17, 2015 at 10:26 PM, Jim Bromer via AGI <[email protected]> >>>>>> wrote: >>>>>>> I started wondering about how a good Satisfiability model might be >>>>>>> used with AGI. >>>>>> It wouldn't because the hard problems in AI like vision and language >>>>>> are not NP-hard. The more useful application would be breaking nearly >>>>>> all forms of cryptography. (One time pad would still be secure). >>>>>> -- Matt Mahoney >>>>> I seriously doubt the premise that the hard problems like vision and >>>>> language in AI are not NP-hard. >>>> NP-hard means NP-complete or harder. NP-complete means that a solution >>>> would solve any problem in NP. NP is the class of problems whose >>>> answers can be verified in time that is a polynomial function of the >>>> input size. P is the class of problems that can be solved in >>>> polynomial time. It is widely believed by everyone except Jim Bromer >>>> that P != NP. This belief is not because of any proof, but because >>>> thousands of other people like Jim Bromer who believed P = NP failed >>>> to find polynomial time solutions to any NP-complete problems after >>>> years of effort until they were convinced they would be better off if >>>> they gave up. The time it takes to give up is inversely proportional >>>> to the person's efforts into studying the math and researching the >>>> work of others instead of repeating their mistakes. >>>> >>>>> My (admittedly limited) experience >>>>> with visual AI ran up against NP-Hard solutions that I thought would >>>>> work. >>>> Vision is a pattern recognition problem. You input a picture of a cat >>>> and output a label like "cat". It is not NP-complete because (1) >>>> experimentally, the problem scales polynomially with input size and >>>> (2) the time to verify that a label like "cat" is correct is about the >>>> same as the time it takes to label the image. Thus, the problem is in >>>> P and would not benefit even if P = NP. >>>> >>>>> And since language could be considered to be a form of >>>>> cryptography then your conjunction of cases (not language but >>>>> cryptography) does not look really strong. >>>> No, language is also a pattern recognition problem. >>>> >>>>> (Visual processing also >>>>> might be considered to be a form of cryptography and indeed it is used >>>>> as such in captchas.) >>>> Cryptography depends on the existence of one-way functions: given >>>> function f and output f(x), you can't find input x any faster than >>>> trying all possible values and comparing the outputs. If P = NP, then >>>> one-way functions would not exist. You could build a circuit that >>>> computes f and compares the output. Then set the bits of x one at a >>>> time and ask your polynomial SAT solver if a solution exists. If not, >>>> flip the bit before going to the next bit. >>>> >>>> You could argue that a captcha is a one way function. It is easy to >>>> convert text to an image, but hard to convert it back. But it is >>>> polynomially hard, not exponentially hard. Adding one bit to the image >>>> doesn't double the solution time, like adding one bit to an encryption >>>> key would. >>>> >>>> -- >>>> -- Matt Mahoney, [email protected] >>>> >>>> >>>> ------------------------------------------- >>>> AGI >>>> Archives: https://www.listbox.com/member/archive/303/=now >>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5 >>>> Modify Your Subscription: https://www.listbox.com/member/?& >>>> Powered by Listbox: http://www.listbox.com >>> >>> ------------------------------------------- >>> AGI >>> Archives: https://www.listbox.com/member/archive/303/=now >>> RSS Feed: https://www.listbox.com/member/archive/rss/303/3701026-786a0853 >>> Modify Your Subscription: https://www.listbox.com/member/?& >>> Powered by Listbox: http://www.listbox.com >> >> >> -- >> -- Matt Mahoney, [email protected] >> >> >> ------------------------------------------- >> AGI >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5 >> Modify Your Subscription: https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/2997756-fc0b9b09 > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
