This sounds good to me. I am much more drawn to topic #1. Topic #2 I have seen discussed recursively and in dozens of variants multiple places. The only thing I will add to Topic #2 is that I very seriously doubt current human intelligence individually or collectively is sufficient to address or meaningfully resolve or even crisply articulate such questions. Much more is accomplished by actually "looking into the horse's mouth" than philosophizing endlessly.

- samantha


Ben Goertzel wrote:

Hi all,

I have been thinking a bit about the nature of conversations on this list.

It seems to me there are two types of conversations here:

1)
Discussions of how to design or engineer AGI systems, using current computers, according to designs that can feasibly be implemented by moderately-sized groups of people

2)
Discussions about whether the above is even possible -- or whether it is impossible because of weird physics, or poorly-defined special characteristics of human creativity, or the so-called "complex systems problem", or because AGI intrinsically requires billions of people and quadrillions of dollars, or whatever

Personally I am pretty bored with all the conversations of type 2.

It's not that I consider them useless discussions in a grand sense ... certainly, they are valid topics for intellectual inquiry. But, to do anything real, you have to make **some** decisions about what approach to take, and I've decided long ago to take an approach of trying to engineer an AGI system.

Now, if someone had a solid argument as to why engineering an AGI system is impossible, that would be important. But that never seems to be the case. Rather, what we hear are long discussions of peoples' intuitions and opinions in this regard. People are welcome to their own intuitions and opinions, but I get really bored scanning through all these intuitions about why AGI is impossible.

One possibility would be to more narrowly focus this list, specifically on **how to make AGI work**.

If this re-focusing were done, then philosophical arguments about the impossibility of engineering AGI in the near term would be judged **off topic** by definition of the list purpose.

Potentially, there could be another list, something like "agi-philosophy", devoted to philosophical and weird-physics and other discussions about whether AGI is possible or not. I am not sure whether I feel like running that other list ... and even if I ran it, I might not bother to read it very often. I'm interested in new, substantial ideas related to the in-principle possibility of AGI, but not interested at all in endless philosophical arguments over various peoples' intuitions in this regard.

One fear I have is that people who are actually interested in building AGI, could be scared away from this list because of the large volume of anti-AGI philosophical discussion. Which, I add, almost never has any new content, and mainly just repeats well-known anti-AGI arguments (Penrose-like physics arguments ... "mind is too complex to engineer, it has to be evolved" ... "no one has built an AGI yet therefore it will never be done" ... etc.)

What are your thoughts on this?

-- Ben




On Wed, Oct 15, 2008 at 10:49 AM, Jim Bromer <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>> wrote:

    On Wed, Oct 15, 2008 at 10:14 AM, Ben Goertzel <[EMAIL PROTECTED]
    <mailto:[EMAIL PROTECTED]>> wrote:
    >
    > Actually, I think COMP=false is a perfectly valid subject for
    discussion on
    > this list.
    >
    > However, I don't think discussions of the form "I have all the
    answers, but
    > they're top-secret and I'm not telling you, hahaha" are
    particularly useful.
    >
    > So, speaking as a list participant, it seems to me this thread
    has probably
    > met its natural end, with this reference to proprietary
    weird-physics IP.
    >
    > However, speaking as list moderator, I don't find this thread so
    off-topic
    > or unpleasant as to formally kill the thread.
    >
    > -- Ben

    If someone doesn't want to get into a conversation with Colin about
    whatever it is that he is saying, then they should just exercise some
    self-control and refrain from doing so.

    I think Colin's ideas are pretty far out there. But that does not mean
    that he has never said anything that might be useful.

    My offbeat topic, that I believe that the Lord may have given me some
    direction about a novel approach to logical satisfiability that I am
    working on, but I don't want to discuss the details about the
    algorithms until I have gotten a chance to see if they work or not,
    was never intended to be a discussion about the theory itself.  I
    wanted to have a discussion about whether or not a good SAT solution
    would have a significant influence on AGI, and whether or not the
    unlikely discovery of an unexpected breakthrough on SAT would serve as
    rational evidence in support of the theory that the Lord helped me
    with the theory.

    Although I am skeptical about what I think Colin is claiming, there is
    an obvious parallel between his case and mine.  There are relevant
    issues which he wants to discuss even though his central claim seems
    to private, and these relevant issues may be interesting.

    Colin's unusual reference to some solid path which cannot be yet
    discussed is annoying partly because it so obviously unfounded.  If he
    had the proof (or a method), then why isn't he writing it up (or
    working it out).  A similar argument was made against me by the way,
    but the difference was that I never said that I had the proof or
    method.  (I did say that you should get used to a polynomial time
    solution to SAT but I never said that I had a working algorithm.)

    My point is that even though people may annoy you with what seems like
    unsubstantiated claims, that does not disqualify everything they have
    said. That rule could so easily be applied to anyone who posts on that
    list.

    Jim Bromer


    -------------------------------------------
    agi
    Archives: https://www.listbox.com/member/archive/303/=now
    RSS Feed: https://www.listbox.com/member/archive/rss/303/
    Modify Your Subscription: https://www.listbox.com/member/?&;
    <https://www.listbox.com/member/?&;>
    Powered by Listbox: http://www.listbox.com




--
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>

"Nothing will ever be attempted if all possible objections must be first overcome " - Dr Samuel Johnson


------------------------------------------------------------------------
*agi* | Archives <https://www.listbox.com/member/archive/303/=now> <https://www.listbox.com/member/archive/rss/303/> | Modify <https://www.listbox.com/member/?&;> Your Subscription [Powered by Listbox] <http://www.listbox.com>




-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34
Powered by Listbox: http://www.listbox.com

Reply via email to