On Fri, Dec 28, 2012 at 6:09 PM, Logan Streondj <[email protected]> wrote:

> It's good to get the ideas into something more tangible like programs.
>

Once you start to create an actual program your original options close in
on your plans. One of the problems in developing a viable AGI program is
that it has to be kept simple.  You might have different methods that are
called in for separate cases but that does not always work so well.


>
> If you find that after whichever period of time you aren't getting much
> anywhere with your chosen route, perhaps you'll choose to contribute to
> another AGI project, perhaps my own.
> I did a version release today, now have support for primitive variables
> :-). By next version release quite possibly will be able to do factorial or
> some other simple procedures.
> And likely by next year will have English grammar,
> allowing for easier verification by others with smaller learning curve
>

What does that mean?  How does it support primitive variables.  And how
would you change your plans if something did not work.  For instance, if it
did not have English grammar how would that affect your concepts about
AGI?  The most important thing is to identify problems that can be solved
and problems that you don't have an answer for.  After working on logical
satisfiability I have come to the conclusion that I don't have a solution
to logical complexity.  So then in order to make my AGI program work it
would have to work by finding a way to overcome the problem of logical
complexity by some other means. It would have to acquire a great deal of
information by serendipity and then make 'intuitive' guesses about
relations that can only be structured through correlation and the
recognition that if process X could be applied to situation Y then it
suggest a path toward finding a solution.

But what if my ideas did not work?  Then it would tell me that I had been
making some mistake.  If I could find good candidates that kept the program
from working then I should be able to test them pretty quickly.  Perhaps
basic correlation is not good enough.  Perhaps the program has to rely on
some kind of enhanced correlation where there are numerous reasons to
believe the process X *can* be applied to situation Y and that it *will*
lead to a path toward a solution.
Jim Bromer




On Fri, Dec 28, 2012 at 6:09 PM, Logan Streondj <[email protected]> wrote:

> Hey I'm just offering you support to do some real coding.
> It's good to get the ideas into something more tangible like programs.
>
> If you find that after whichever period of time you aren't getting much
> anywhere with your chosen route, perhaps you'll choose to contribute to
> another AGI project, perhaps my own.
>
> I did a version release today, now have support for primitive variables
> :-).  By next version release quite possibly will be able to do  factorial
> or some other simple procedures.
> And likely by next year will have English grammar,
> allowing for easier verification by others with smaller learning curve.
>
> I'm programming in Assembly, but it is quite simple,
> only 16 assembly commands used, all register-machine,
> makes it easy to port and that kinda stuff.
>
> You would certainly have the capacity to improve upon current AGI
> programs,  can look at the current roadmap and see where your ideas might
> fit in
> https://sourceforge.net/p/rpoku/code/ci/dc0d7886965d5cab645a4d5a220391b316c7c388/tree/roadmap.txt?format=raw
>
>
> On Fri, Dec 28, 2012 at 10:53 AM, Jim Bromer <[email protected]> wrote:
>
>> The most important case would be the one where it does show some
>> capability of learning a crude simplistic language but where it either
>> lacks subtlety or where it shows a wide variation of depth.  In some cases,
>> for example, it might seem to be working but then it just cannot continue
>> to learn new things about a particular subject or where other subjects
>> which are comparably as easy seem to be totally beyond it. This is along
>> the lines of how other AI projects have fared.  Let's say that my project
>> did turn out like this.  Then in order to show that it was a valid concept
>> I would have to advance the program so that it was able to go further than
>> it had.  The thing is that although the various AI methods are able to do
>> some tasks better than others they all fail at a level below what we need
>> to see in order to compare them to children.  So being human like is not
>> the immediate goal, and being really smart is not the immediate goal. But I
>> would need to show that I could improve on contemporary AGI programs in
>> order to demonstrate that my ideas were workable and since my program would
>> be limited I would need to show that some improvements could be made to my
>> program.
>>
>> Jim Bromer
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/5037279-a88c7a6d> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>
>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10561250-470149cf> |
> Modify<https://www.listbox.com/member/?&;>Your Subscription
> <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to