I think I understand what you are getting at, and it makes a lot of
sense. You and Aaron have convinced me that I should spend more time
working on my AI / AGI project but unfortunately I still do not seem
to have the time to work on it.

I think I do have some good ideas about things like artificial
imagination which is important.  And curiosity is something that I
always felt was easy.
Jim Bromer


On Sun, Dec 6, 2015 at 3:41 PM, Stanley Nilsen <[email protected]> wrote:
> Thanks for giving this some thought Jim.  I'm going out of town for a few
> days, so don't consider silence to be a loss of interest.
>
> One of your comments was:
>
> "But the program has to be able to develop its
> own strategies to 'evaluate' some things because that is a good
> strategy for a computer program to use - in some cases. And the
> usefulness of logical 'evaluation' implies that some strategy for
> evaluating conceptual relationships other than simple numerical
> methods would also be a good strategy to use."
>
> ---------------------
> My problem with the program developing "it's own strategy to evaluate..." is
> that strategy is not a strength of a child. Somehow children acquire the
> ability to put 2 and 2 together, but we haven't discovered how to get a
> machine to do it.  What's the machine equivalent of curiosity?  I'm not
> convinced that we have an adequate "big" picture to see how the pieces will
> eventually fit together.
>
> The big picture looks kind of like "design and make a system that works,
> even if one needs to, substitute human effort for some of the components."
> Then, when the system is in place, determine how to remove more and more of
> the human element.  Eventually one is left with a system that may interface
> with humans but only as though using them as a resource.
>
> By the way, I think a text only approach is a good start.  I'm interested in
> looking at the use of words as a way to convey "benefit."  Initial design is
> interesting because there are so many words and phrases to choose from.  I
> get it that this sounds like a chat bot, but for me it's a way of
> experimenting with the idea of a benefit driven system.
>
> Stan
>
>
>
> On 12/06/2015 09:17 AM, Jim Bromer wrote:
>>
>> You might be able to think of ways to benefit the poor but you would
>> have a lot of trouble to implement them. You might be able to help a
>> few people but if you are like most of the rest of us that would be
>> it.
>>
>> So you think that there are a lot of opportunities to use basic
>> implementation strategies to get the AI/AGI program to do something
>> that would be beneficial in some way? But the only problem that you
>> foresee is the coding? But why would that be difficult? For example, I
>> think that I could develop a prototype of an AGI program using text
>> only. If you start with something like that then it would be simple to
>> get started because you can find code that contains the basic forms
>> for text IO. The problem that I am having is that even when I strip
>> the plan down to what I think would be a minimum for a simple database
>> management program (of my own design) it still cannot be done on the
>> little time I have to code, and without any reason to believe that I
>> could get past something that would not work too well I don't have
>> much commitment to get going on it.
>>
>> You said:
>> "Values (rules about values) come into play as the AGI picks the next
>> thing to do.  But, we already know that early AGI doesn't have a
>> "values" structure to refer to.  To program one is really not much of
>> an option - it is too complex to "calculate" what the value of
>> something is.  To test the validity of my statement that it is too
>> complex to calculate, try it. Imagine that you are writing this into
>> code!"
>>
>> I have tried to imagine writing that into code! (Why wouldn't I have
>> tried to imagine that?) But the program has to be able to develop its
>> own strategies to 'evaluate' some things because that is a good
>> strategy for a computer program to use - in some cases. And the
>> usefulness of logical 'evaluation' implies that some strategy for
>> evaluating conceptual relationships other than simple numerical
>> methods would also be a good strategy to use. But this would be
>> complicated. I think the opportunities that you mentioned would be
>> difficult to code as well - if you wanted to avoid getting bogged down
>> in code that is good for narrow-AI. The problem is that once you make
>> the commitment to do something that is effectively narrow-AI then
>> there are all sorts of enticing shortcuts that become available but
>> that you really need to keep to a minimum.
>>
>> Using a text-only program that has to start so that it can only act on
>> the simple 'opportunities' (or 'low hanging fruit') of text (and
>> conversation of course) is where I would start. But it should be clear
>> that I don't want to take all the shortcuts that sort of situation
>> would offer. So I want my program to 'look' for opportunities on its
>> own so to speak. It may not be possible for a program to do that at a
>> very sophisticated level from our point of view, but we know that
>> computer programs are good at some things that we are not so good at.
>> So, my point of view is that the program should be able to pick up all
>> sorts of patterns (opportunities) that we would miss so that is where
>> I want to start at. Having thought about that I concluded that it
>> would have to be looking at the recombination of all sorts of odd
>> kinds of data in order to find a few combinations that might be
>> useful.
>> Jim Bromer
>>
>>
>
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-653794b5
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to