Matt
I appreciate the conversations we have had about the N=?NP question.
My interest in the subject concerns the question of how significant
improvements in SAT solutions could affect development of AGI
programs. The reason I have not been interested in joining discussion
groups of N=?NP is just because my primary interest is in semi-strong
AI. I think it is obvious that significant improvements in SAT
solutions would  produce significant advances in AGI.

If you think our conversations about this have gone as far as they can
go, then ok. But I think there is a lot of room for improvement in SAT
Solutions and I think that those solutions will have a dramatic impact
in the development of AGI programs. I have tried to describe some
possible scenarios of how this might be. Of course, my tortured use of
language to describe these --virtual multi level semi-permeable
bounded local logical systems in a coherentist model of knowledge--
may not be immediately obvious.

I can't casually describe an algorithm where P=NP would exponentially
speed up visual processing because I am not an expert in the field so
it would take a great deal of time just to try to explain my view
point on the subject. (However that is an interesting question and I
will think about it.) Sufficient to say that I believe that the
premise that a neural network would be similar to a system in the
neural cortex is not sound. Visual processing is not entirely local so
even if someone actually knew how the visual cortex worked and could
convince me that he did, I still would not accept the argument that a
snippet of the visual cortex would adequately describe how visual
analysis and recognition works.

But I will take a minute to talk about the background of my view on
the subject. Suppose that you were trying to find an object on the
ground in a satellite image. Suppose that at a low resolution it was
impossible to find a small object. As the resolution increased the
object might start to appear but it would still be difficult to spot
it. As the resolution increased more you might be able to find the
object if you knew what you were looking for. From that argument you
might say that visual recognition is polynomial. Well if the ground
area that was covered by the picture remained the same that would mean
that you would have to search through more detail to find the object
as the resolution increases. Suppose the resolution increased way
beyond the scale (of the object) that you were familiar with. Now, not
only would the 'perceived area' that needs to be searched increase
beyond your ability to cope with it, you might still not be able to
recognize the object even if you looked right at it regardless of the
overabundance of detail. Recognition is dependent on being able to
make out enough detail of the object but it also is dependent on
ignoring the variants and other objects that act like noise in the
image (relevant to the recognition task). Even so, recognition has to
use previous visual experiences with objects like the one you are
looking for to compare the various candidates against. I think it is
safe to say that, in this type of example, as the increases in
resolution makes the image size increase polynomially that the
difficulty of finding an object can - in the worse cases - increase
exponentially (or worse). The object is still there but you won't find
it examining one pixel at a time, you have to see the object in terms
numerous pixels and typically, you have to weigh how various groups of
pixels (of candidates) relates to the other stuff in the image.
Jim Bromer
Jim Bromer


On Fri, Mar 6, 2015 at 10:04 AM, Matt Mahoney via AGI <[email protected]> wrote:
> Jim, can you describe an algorithm where P = NP would exponentially
> speed up visual processing? My understanding is that the most advanced
> vision algorithms use deep neural networks with a structure similar to
> the visual cortex. In general, neural network size (in synapses)
> should be proportional to the training set information content. Thus,
> training time is O(n^2).
>
> On Thu, Mar 5, 2015 at 10:01 PM, Jim Bromer via AGI <[email protected]> wrote:
>>  Matt said:
>> Vision is a pattern recognition problem. You input a picture of a cat
>> and output a label like "cat". It is not NP-complete because (1)
>> experimentally, the problem scales polynomially with input size and
>> (2) the time to verify that a label like "cat" is correct is about the
>> same as the time it takes to label the image. Thus, the problem is in
>> P and would not benefit even if P = NP.
>> -------------------------------------------------
>> This is a truly insipid response. You have taken one narrow situation
>> and used it in an over-generalization of a kind of AGI problem. "The
>> problem scales polynomially with input size? The point that I made is
>> that the general analysis of imagery is presumably as bad or worse
>> than NP (in the lexicon of the day). What I mean is that there is
>> sufficient evidence that AGI is, in the worse case, at least
>> exponentially difficult and that makes it worthwhile to examine why
>> that may be. One reason, the reason I gave, is that the easiest
>> methods to make a methodical and thorough analysis of the relations
>> between associated pixels would be those that are (literally) in NP.
>> The implied case of scaling a particular picture and arguing that it
>> would scale polynomially with input size is analogous to saying that
>> converting an unrestricted Boolean Satisfiability problem to 3-SAT
>> scales polynomially (and that somehow proves that unrestricted SAT
>> scales polynomially). It is pretty obvious that you have little
>> experience with visual data.
>>
>> This is an example of a blatant overgeneralization being declared as
>> if  it were a factual statement. I can't casually explain why visual
>> analysis is at least exponentially difficult because I am not enough
>> of an expert to be that familiar with all the problems. However, I am
>> confident that there is no overwhelming evidence to suggest that, in
>> general, it is less difficult.
>> Jim Bromer
>>
>>
>> On Thu, Mar 5, 2015 at 1:04 PM, Matt Mahoney via AGI <[email protected]> 
>> wrote:
>>> On Wed, Mar 4, 2015 at 2:51 AM, Jim Bromer <[email protected]> wrote:
>>>>  On Tue, Feb 17, 2015 at 11:52 PM, Matt Mahoney via AGI <[email protected]> 
>>>> wrote:
>>>>> On Tue, Feb 17, 2015 at 10:26 PM, Jim Bromer via AGI <[email protected]> 
>>>>> wrote:
>>>>>> I started wondering about how a good Satisfiability model might be
>>>>>> used with AGI.
>>>>>
>>>>> It wouldn't because the hard problems in AI like vision and language
>>>>> are not NP-hard. The more useful application would be breaking nearly
>>>>> all forms of cryptography. (One time pad would still be secure).
>>>>> -- Matt Mahoney
>>>>
>>>> I seriously doubt the premise that the hard problems like vision and
>>>> language in AI are not NP-hard.
>>>
>>> NP-hard means NP-complete or harder. NP-complete means that a solution
>>> would solve any problem in NP. NP is the class of problems whose
>>> answers can be verified in time that is a polynomial function of the
>>> input size. P is the class of problems that can be solved in
>>> polynomial time. It is widely believed by everyone except Jim Bromer
>>> that P != NP. This belief is not because of any proof, but because
>>> thousands of other people like Jim Bromer who believed P = NP failed
>>> to find polynomial time solutions to any NP-complete problems after
>>> years of effort until they were convinced they would be better off if
>>> they gave up. The time it takes to give up is inversely proportional
>>> to the person's efforts into studying the math and researching the
>>> work of others instead of repeating their mistakes.
>>>
>>>> My (admittedly limited) experience
>>>> with visual AI ran up against NP-Hard solutions that I thought would
>>>> work.
>>>
>>> Vision is a pattern recognition problem. You input a picture of a cat
>>> and output a label like "cat". It is not NP-complete because (1)
>>> experimentally, the problem scales polynomially with input size and
>>> (2) the time to verify that a label like "cat" is correct is about the
>>> same as the time it takes to label the image. Thus, the problem is in
>>> P and would not benefit even if P = NP.
>>>
>>>> And since language could be considered to be a form of
>>>> cryptography then your conjunction of cases (not language but
>>>> cryptography) does not look really strong.
>>>
>>> No, language is also a pattern recognition problem.
>>>
>>>> (Visual processing also
>>>> might be considered to be a form of cryptography and indeed it is used
>>>> as such in captchas.)
>>>
>>> Cryptography depends on the existence of one-way functions: given
>>> function f and output f(x), you can't find input x any faster than
>>> trying all possible values and comparing the outputs. If P = NP, then
>>> one-way functions would not exist. You could build a circuit that
>>> computes f and compares the output. Then set the bits of x one at a
>>> time and ask your polynomial SAT solver if a solution exists. If not,
>>> flip the bit before going to the next bit.
>>>
>>> You could argue that a captcha is a one way function. It is easy to
>>> convert text to an image, but hard to convert it back. But it is
>>> polynomially hard, not exponentially hard. Adding one bit to the image
>>> doesn't double the solution time, like adding one bit to an encryption
>>> key would.
>>>
>>> --
>>> -- Matt Mahoney, [email protected]
>>>
>>>
>>> -------------------------------------------
>>> AGI
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>
>>
>> -------------------------------------------
>> AGI
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/3701026-786a0853
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>
>
>
> --
> -- Matt Mahoney, [email protected]
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to