The most effective image recognition systems we have today, including
semantic understanding of scenes, are all NN variants to the best of my
knowledge.

-s

On 03/08/2015 08:51 PM, Jim Bromer via AGI wrote:
>  > Jim, can you describe an algorithm where P = NP would exponentially
>> speed up visual processing?
> The complexity of trying to find how each pixel (or tiny area) relates
> to other pixels is not a true Satisfiability problem until a (more
> classic-like) image analysis algorithm gathers some information. For
> example, is some area a part of the background? That is not an easy
> question because the 'background' may be quite diverse. Furthermore,
> the desired 'subjects' of an image may not be -perceived- before image
> analysis gets going. Once an algorithm starts picking up information
> than the variations of possibilities starts forming a true
> Satisfiaibilty problem although it may not be expressed in simple
> Boolean relations.
>
> While most image analysis would take place across a field of data that
> does not mean that all image analysis are essentially neural networks.
>
> I don't want to dismiss deep learning neural networks just because
> they have not achieved even shallow AGI. But look at character
> recognition. Alphabet characters have flat distinct shapes. Although
> they may vary widely, one might still design a classical algorithm
> that defines how near a subject character is to a set of training
> characters in the terms of vectors (and weighted reasoning) or
> something like that. The attempt to use of vectors with 'ideas' or
> semantic objects has been  inadequate because the domains of 'ideas'
> do not all fit into a domain of vector space. Space and much of
> physics have benefited from a great deal of effective mathematical
> analysis over the centuries. So computers are great at predicting the
> weather because the averages of history (of different measures of
> events) can be combined with knowledge of the causes of the weather
> and presto - you have some great weather predictions. But if you
> wanted to push deep anns what are you going to find? You are going to
> find that you have to push up against the complexity barriers.
> Although the Boolean relations between areas of an image in an ann may
> be hidden, they are never the less creating complexity barriers. Anns
> work so well because some problems can be effectively solved by using
> multiple metric approximations.
>
> But, let's suppose that AGI will one day be accomplished using metrics
> - in other words weighted reasoning, probability and so on. I realize
> that it is a real possibility. That means that image analysis
> algorithms will be able to recognize anything a human being could
> recognize. Here the problem is not a question of taking tens or
> hundreds of flat land characteristics for tens of characters and
> coming up with a method that effectively measures how close a subject
> character is to different kinds of metrics derived from the training
> characters and then picks the closest matches, but of taking thousands
> of characteristics from thousands of objects to derive estimates of
> what is pictured. In this case the problem is combinatorially more
> complex just because there are so many more possible subjects and
> variants of those subjects. It should be clear that this is a true
> satisfiability problem even though Anns don't deal directly with the
> satisfiability issue because the acquired weights are so heavily
> distributed. The complexity barriers are effectively satisfiability
> problems even though the distinct relations may be hidden. I hope this
> makes some sense because I really don't know that much about deep
> learning Anns and I haven't really done that much image analysis.
>
> One other thing. As you know, a neural network is not the only kind of
> algorithm that can learn. So it is pretty easy to imagine an algorithm
> that is based on supervised learning to develop discreet relations
> that form networks. The relations would of course include weights and
> so on. One of the reasons I would like to develop some better SAT
> methods is that I then could develop AI models using simple concepts
> and build on it. Right now the complexity barriers are so low and so
> pervasive that you can't get any real traction unless the problem can
> be solved using weighted approximations. So weighted reasoning is way
> ahead right now but that may not always be the case. My goal is not to
> achieve detailed precision but to get some basic traction by starting
> with something simple and improving it.
>
> Jim Bromer
>
>
> On Fri, Mar 6, 2015 at 10:04 AM, Matt Mahoney via AGI <[email protected]> 
> wrote:
>> Jim, can you describe an algorithm where P = NP would exponentially
>> speed up visual processing? My understanding is that the most advanced
>> vision algorithms use deep neural networks with a structure similar to
>> the visual cortex. In general, neural network size (in synapses)
>> should be proportional to the training set information content. Thus,
>> training time is O(n^2).
>>
>> On Thu, Mar 5, 2015 at 10:01 PM, Jim Bromer via AGI <[email protected]> wrote:
>>>  Matt said:
>>> Vision is a pattern recognition problem. You input a picture of a cat
>>> and output a label like "cat". It is not NP-complete because (1)
>>> experimentally, the problem scales polynomially with input size and
>>> (2) the time to verify that a label like "cat" is correct is about the
>>> same as the time it takes to label the image. Thus, the problem is in
>>> P and would not benefit even if P = NP.
>>> -------------------------------------------------
>>> This is a truly insipid response. You have taken one narrow situation
>>> and used it in an over-generalization of a kind of AGI problem. "The
>>> problem scales polynomially with input size? The point that I made is
>>> that the general analysis of imagery is presumably as bad or worse
>>> than NP (in the lexicon of the day). What I mean is that there is
>>> sufficient evidence that AGI is, in the worse case, at least
>>> exponentially difficult and that makes it worthwhile to examine why
>>> that may be. One reason, the reason I gave, is that the easiest
>>> methods to make a methodical and thorough analysis of the relations
>>> between associated pixels would be those that are (literally) in NP.
>>> The implied case of scaling a particular picture and arguing that it
>>> would scale polynomially with input size is analogous to saying that
>>> converting an unrestricted Boolean Satisfiability problem to 3-SAT
>>> scales polynomially (and that somehow proves that unrestricted SAT
>>> scales polynomially). It is pretty obvious that you have little
>>> experience with visual data.
>>>
>>> This is an example of a blatant overgeneralization being declared as
>>> if  it were a factual statement. I can't casually explain why visual
>>> analysis is at least exponentially difficult because I am not enough
>>> of an expert to be that familiar with all the problems. However, I am
>>> confident that there is no overwhelming evidence to suggest that, in
>>> general, it is less difficult.
>>> Jim Bromer
>>>
>>>
>>> On Thu, Mar 5, 2015 at 1:04 PM, Matt Mahoney via AGI <[email protected]> 
>>> wrote:
>>>> On Wed, Mar 4, 2015 at 2:51 AM, Jim Bromer <[email protected]> wrote:
>>>>>  On Tue, Feb 17, 2015 at 11:52 PM, Matt Mahoney via AGI 
>>>>> <[email protected]> wrote:
>>>>>> On Tue, Feb 17, 2015 at 10:26 PM, Jim Bromer via AGI <[email protected]> 
>>>>>> wrote:
>>>>>>> I started wondering about how a good Satisfiability model might be
>>>>>>> used with AGI.
>>>>>> It wouldn't because the hard problems in AI like vision and language
>>>>>> are not NP-hard. The more useful application would be breaking nearly
>>>>>> all forms of cryptography. (One time pad would still be secure).
>>>>>> -- Matt Mahoney
>>>>> I seriously doubt the premise that the hard problems like vision and
>>>>> language in AI are not NP-hard.
>>>> NP-hard means NP-complete or harder. NP-complete means that a solution
>>>> would solve any problem in NP. NP is the class of problems whose
>>>> answers can be verified in time that is a polynomial function of the
>>>> input size. P is the class of problems that can be solved in
>>>> polynomial time. It is widely believed by everyone except Jim Bromer
>>>> that P != NP. This belief is not because of any proof, but because
>>>> thousands of other people like Jim Bromer who believed P = NP failed
>>>> to find polynomial time solutions to any NP-complete problems after
>>>> years of effort until they were convinced they would be better off if
>>>> they gave up. The time it takes to give up is inversely proportional
>>>> to the person's efforts into studying the math and researching the
>>>> work of others instead of repeating their mistakes.
>>>>
>>>>> My (admittedly limited) experience
>>>>> with visual AI ran up against NP-Hard solutions that I thought would
>>>>> work.
>>>> Vision is a pattern recognition problem. You input a picture of a cat
>>>> and output a label like "cat". It is not NP-complete because (1)
>>>> experimentally, the problem scales polynomially with input size and
>>>> (2) the time to verify that a label like "cat" is correct is about the
>>>> same as the time it takes to label the image. Thus, the problem is in
>>>> P and would not benefit even if P = NP.
>>>>
>>>>> And since language could be considered to be a form of
>>>>> cryptography then your conjunction of cases (not language but
>>>>> cryptography) does not look really strong.
>>>> No, language is also a pattern recognition problem.
>>>>
>>>>> (Visual processing also
>>>>> might be considered to be a form of cryptography and indeed it is used
>>>>> as such in captchas.)
>>>> Cryptography depends on the existence of one-way functions: given
>>>> function f and output f(x), you can't find input x any faster than
>>>> trying all possible values and comparing the outputs. If P = NP, then
>>>> one-way functions would not exist. You could build a circuit that
>>>> computes f and compares the output. Then set the bits of x one at a
>>>> time and ask your polynomial SAT solver if a solution exists. If not,
>>>> flip the bit before going to the next bit.
>>>>
>>>> You could argue that a captcha is a one way function. It is easy to
>>>> convert text to an image, but hard to convert it back. But it is
>>>> polynomially hard, not exponentially hard. Adding one bit to the image
>>>> doesn't double the solution time, like adding one bit to an encryption
>>>> key would.
>>>>
>>>> --
>>>> -- Matt Mahoney, [email protected]
>>>>
>>>>
>>>> -------------------------------------------
>>>> AGI
>>>> Archives: https://www.listbox.com/member/archive/303/=now
>>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
>>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>>> Powered by Listbox: http://www.listbox.com
>>>
>>> -------------------------------------------
>>> AGI
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/3701026-786a0853
>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>
>>
>> --
>> -- Matt Mahoney, [email protected]
>>
>>
>> -------------------------------------------
>> AGI
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/2997756-fc0b9b09
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to