One month has gone by and the progress I have made on my AGi project is so
slow that it is obvious to me that I will not be able to get it working in
a year or two unless I do a lot better.  So using a predicted date in a
theory about the development of my project has allowed me to get a better
sense of how I am doing even though there are no true tests to benchmark
this.

Expectations may be implicitly or explicitly associated with a great deal
of knowledge but I just do not feel that they are essential objectives of
intelligence.  Making a thorough philosophical analysis of how expectation
plays in different theories about intelligence/AI/AGI is probably not the
best use of my time, but let me say that there are some subtleties involved.

It is very useful to find observable objectives that can be used to
establish some sense of the effectiveness and proper usage of a theory, and
these observable objectives can even be associated with more elusive
objectives but the more elusive parts of combined objectives have to be
used wisely.  So my *feelings* of how much progress I have made on my AGi
project are subjective "objectives" but as long as I am honest and willing
to put some thought into it I can interpret them wisely.  I am 1/12 of the
way to my "deadline," can I say that my program is 1/12 of the way to being
intelligent?  No.

The fact that I am doing more programming and that I have a better plan
than I did before are encouraging signs.  And while I haven't discovered
anything about AGI I have discovered something.  When I wanted to try to
run a simple initial test of intelligence I realized that it was beyond me
because intelligence requires a great deal of integrated knowledge to serve
as the potential background for a simple test.  So I realized that I will
not be able to try exactly what I wanted to try.  So instead I will need to
create some novel initial tests (of AGi) by combining what I discover can
be done with a number of simple algorithms that I am using and then
applying my imagination to see how I might use those algorithms to gain
some kind of artificial insight.  So this may be a trivial insight about
developing an AGI project but I believe that the fact that I have formally
recognized it will make it more likely that I will be able to develop good
initial AGi tests than I might have been otherwise.  I will have a better
sense of well this new plan will work by the start of the next month.

Jim Bromer



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to