Re: [agi] Free AI Courses at Stanford

2008-09-18 Thread Kingma, D.P.
Hi List, Also interesting to some of you may be VideoLectures.net, which offers lots of interesting lectures. Although not all are of Stanford quality, still I found many interesting lectures by respected lecturers. And there are LOTS (625 at the moment) of lectures about Machine Learning... :)

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread David Hart
On Thu, Sep 18, 2008 at 3:26 PM, Linas Vepstas [EMAIL PROTECTED]wrote: I agree that the topic is worth careful consideration. Sacrificing the 'free as in freedom' aspect of AGPL-licensed OpenCog for reasons of AGI safety and/or the prevention of abuse may indeed be necessary one day.

Re: [agi] uncertain logic criteria

2008-09-18 Thread Pei Wang
On Wed, Sep 17, 2008 at 10:54 PM, Abram Demski [EMAIL PROTECTED] wrote: Pei, You are right, that does sound better than quick-and-dirty. And more relevant, because my primary interest here is to get a handle on what normative epistemology should tell us to conclude if we do not have time to

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread Trent Waddington
On Thu, Sep 18, 2008 at 8:08 PM, David Hart [EMAIL PROTECTED] wrote: Original works produced by software as a tool where a human operator is involved at some stage is a different case from original works produced by software exclusively and entirely under its own direction. The latter has no

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread Bob Mottram
2008/9/18 Trent Waddington [EMAIL PROTECTED]: And this is the problem. Although some people have the goal of making an artificial person with all the richness and nuance of a sentient creature with thoughts and feelings and yada yada yada.. some of us are just interested in making more

Re: [agi] Re: [OpenCog] Proprietary_Open_Source

2008-09-18 Thread Matt Mahoney
2008/9/17 JDLaw [EMAIL PROTECTED]: IMHO to all, There is an important morality discussion about how sentient life will be treated that has not received its proper treatment in your discussion groups. I have seen glimpses of this topic, but no real action proposals. How would you feel if

Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Bob Mottram [EMAIL PROTECTED] wrote: And this is the problem. Although some people have the goal of making an artificial person with all the richness and nuance of a sentient creature with thoughts and feelings and yada yada yada.. some of us are just interested

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread Linas Vepstas
2008/9/18 David Hart [EMAIL PROTECTED]: On Thu, Sep 18, 2008 at 3:26 PM, Linas Vepstas [EMAIL PROTECTED] wrote: I agree that the topic is worth careful consideration. Sacrificing the 'free as in freedom' aspect of AGPL-licensed OpenCog for reasons of AGI safety and/or the prevention of

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Ben Goertzel
Lets distinguish between the two major goals of AGI. The first is to automate the economy. The second is to become immortal through uploading. Peculiarly, you are leaving out what to me is by far the most important and interesting goal: The creation of beings far more intelligent than humans

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Steve Richfield
Ben, IMHO... On 9/18/08, Ben Goertzel [EMAIL PROTECTED] wrote: Lets distinguish between the two major goals of AGI. The first is to automate the economy. The second is to become immortal through uploading. Peculiarly, you are leaving out what to me is by far the most important and

[agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Pei Wang
TITLE: Case-by-case Problem Solving (draft) AUTHOR: Pei Wang ABSTRACT: Case-by-case Problem Solving is an approach in which the system solves the current occurrence of a problem instance by taking the available knowledge into consideration, under the restriction of available resources. It is

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Ben Goertzel [EMAIL PROTECTED] wrote: Lets distinguish between the two major goals of AGI. The first is to automate the economy. The second is to become immortal through uploading. Peculiarly, you are leaving out what to me is by far the most important and interesting goal:

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Mike Tintner
Steve:View #2 (mine, stated from your approximate viewpoint) is that simple programs (like Dr. Eliza) have in the past and will in the future do things that people aren't good at. This includes tasks that encroach on intelligence, e.g. modeling complex phonema and refining designs. Steve, In

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread Nathan Cravens
When an AGI writes a book, designs a new manufacturing base, forms a decentralised form of regulation, ect, the copyright and patent system will be futile, because the enclosed material, when deemed useful by another, will access the same information and rewrite it in another form to create a

Re: [agi] self organization

2008-09-18 Thread Bryan Bishop
On Wednesday 17 September 2008, Terren Suydam wrote: I think a similar case could be made for a lot of large open source projects such as Linux itself. However, in this case and others, the software itself is the result of a high-level super goal defined by one or more humans. Even if no

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Charles Hixson
I would go further. Humans have demonstrated that they cannot be trusted in the long term even with the capabilities that we already possess. We are too likely to have ego-centric rulers who make decisions not only for their own short-term benefit, but with an explicit After me the deluge

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Mike Tintner
TITLE: Case-by-case Problem Solving (draft) AUTHOR: Pei Wang ABSTRACT: Case-by-case Problem Solving is an approach in which the system solves the current occurrence of a problem instance by taking the available knowledge into consideration, under the restriction of available resources. It is

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Vladimir Nesov
On Fri, Sep 19, 2008 at 1:31 AM, Trent Waddington [EMAIL PROTECTED] wrote: On Fri, Sep 19, 2008 at 3:36 AM, Matt Mahoney [EMAIL PROTECTED] wrote: Lets distinguish between the two major goals of AGI. The first is to automate the economy. The second is to become immortal through uploading. Umm,

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread David Hart
On Fri, Sep 19, 2008 at 3:53 AM, Linas Vepstas [EMAIL PROTECTED]wrote: Exactly. If opencog were ever to reach the point of popularity where one might consider a change of licensing, it would also be the case that most of the interested parties would *not* be under SIAI control, and thus

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread David Hart
On Thu, Sep 18, 2008 at 9:44 PM, Trent Waddington [EMAIL PROTECTED] wrote: Claiming a copyright and successfully defending that claim are different things. What ways do you envision someone challenging the copyright? Take the hypothetical case of R. Marketroid, who's hardware is on the

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Bryan Bishop
On Thursday 18 September 2008, Mike Tintner wrote: In principle, I'm all for the idea that I think you (and perhaps Bryan) have expressed of a GI Assistant - some program that could be of general assistance to humans dealing with similar problems across many domains. A diagnostics expert,

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Vladimir Nesov [EMAIL PROTECTED] wrote: And to boot, both of you don't really know what you want. What we want has been programmed into our brains by the process of evolution. I am not pretending the outcome will be good. Once we have the technology to have everything we

Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source

2008-09-18 Thread Trent Waddington
On Fri, Sep 19, 2008 at 7:30 AM, David Hart [EMAIL PROTECTED] wrote: Take the hypothetical case of R. Marketroid, who's hardware is on the books as an asset at ACME Marketing LLC and who's programming has been tailered by ACME to suit their needs. Unbeknownst to ACME, RM has decided to write

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Ben Goertzel
Matt M wrote: Peculiarly, you are leaving out what to me is by far the most important and interesting goal: The creation of beings far more intelligent than humans yet benevolent toward humans That's what I mean by an automated economy. Google is already more intelligent than any human

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Trent Waddington [EMAIL PROTECTED] wrote: On Fri, Sep 19, 2008 at 3:36 AM, Matt Mahoney [EMAIL PROTECTED] wrote: Lets distinguish between the two major goals of AGI. The first is to automate the economy. The second is to become immortal through uploading. Umm, who's

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Trent Waddington
On Fri, Sep 19, 2008 at 6:57 AM, Ben Goertzel [EMAIL PROTECTED] wrote: general intelligence at the human level I hear you say these words a lot. I think, by using the word level, you're trying to say something different to general intelligence just like humans have but I'm not sure everyone

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Trent Waddington
On Fri, Sep 19, 2008 at 7:54 AM, Matt Mahoney [EMAIL PROTECTED] wrote: Perhaps there are some applications I haven't thought of? Bahahaha.. Gee, ya think? Trent --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed:

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Ben Goertzel
On Thu, Sep 18, 2008 at 5:42 PM, Mike Tintner [EMAIL PROTECTED]wrote: TITLE: Case-by-case Problem Solving (draft) AUTHOR: Pei Wang But you seem to be reinventing the term for wheel. There is an extensive literature, including AI stuff, on wicked, ill-structured problems, (and even

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Pei Wang [EMAIL PROTECTED] wrote: URL: http://nars.wang.googlepages.com/wang.CaseByCase.pdf I think it would be interesting if you had some experimental results. Could CPS now solve a problem like sort [3 2 4 1] in its current state? If not, how much knowledge does it

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Mike Tintner
Ben, I'm only saying that CPS seems to be loosely equivalent to wicked, ill-structured problem-solving, (the reference to convergent/divergent (or crystallised vs fluid) etc is merely to point out a common distinction in psychology between two kinds of intelligence that Pei wasn't aware of in

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Ben Goertzel
A key point IMO is that: problem-solving that is non-algorithmic (in Pei's sense) at one level (the level of the particular problem being solved) may still be algorithmic at a different level (for instance, NARS itself is a set of algorithms). So, to me, calling NARS problem-solving

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Ben Goertzel
On Thu, Sep 18, 2008 at 9:02 PM, Matt Mahoney [EMAIL PROTECTED] wrote: --- On Thu, 9/18/08, Ben Goertzel [EMAIL PROTECTED] wrote: I believe there is a qualitative difference btw AGI and narrow-AI, so that no tractably small collection of computationally-feasible narrow-AI's (like Google

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Mike Tintner
Ben, Ah well, then I'm confused. And you may be right - I would just like clarification. You see, what you have just said is consistent with my understanding of Pei up till now. He explicitly called his approach in the past nonalgorithmic while acknowledging that others wouldn't consider it

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Ben Goertzel
On Thu, Sep 18, 2008 at 9:17 PM, Ben Goertzel [EMAIL PROTECTED] wrote: Your language is unclear Could you define precisely what you mean by an algorithm Also, could you give an example of a computer program, that can be run on a digital computer, that is not does not embody an algorithm

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Ben Goertzel
Your language is unclear Could you define precisely what you mean by an algorithm Also, could you give an example of a computer program, that can be run on a digital computer, that is not does not embody an algorithm according to your definition? thx ben On Thu, Sep 18, 2008 at 9:15 PM, Mike

Re: [agi] uncertain logic criteria

2008-09-18 Thread YKY (Yan King Yin)
On Thu, Sep 18, 2008 at 4:21 AM, Kingma, D.P. [EMAIL PROTECTED] wrote: Small question... aren't Bbayesian network nodes just _conditionally_ independent: so that set A is only independent from set B when d-separated by some set Z? So please clarify, if possible, what kind of independence you

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Trent Waddington [EMAIL PROTECTED] wrote: On Fri, Sep 19, 2008 at 7:54 AM, Matt Mahoney [EMAIL PROTECTED] wrote: Perhaps there are some applications I haven't thought of? Bahahaha.. Gee, ya think? So perhaps you could name some applications of AGI that don't fall

Re: [agi] Case-by-case Problem Solving PS

2008-09-18 Thread Mike Tintner
Ben, It's hard to resist my interpretation here - that Pei does sound as if he is being truly non-algorithmic. Just look at the opening abstract sentences. (However, I have no wish to be pedantic - I'll accept whatever you guys say you mean). Case-by-case Problem Solving is an approach in

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Matt Mahoney
Actually, CPS doesn't mean solving problems without algorithms. CPS is itself an algorithm, as described on pages 7-8 of Pei's paper. However, as I mentioned, I would be more convinced if there were some experimental results showing that it actually worked. -- Matt Mahoney, [EMAIL PROTECTED]

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread John LaMuth
You have completely left out the human element or friendly-type appeal How about a AGI personal assistant / tutor / PR interface Everyone should have one The market would be virtually unlimited ... John L www.ethicalvalues.com - Original Message - From: Matt Mahoney [EMAIL

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Trent Waddington
On Fri, Sep 19, 2008 at 11:34 AM, Matt Mahoney [EMAIL PROTECTED] wrote: So perhaps you could name some applications of AGI that don't fall into the categories of (1) doing work or (2) augmenting your brain? Perhaps you could list some uses of a computer that don't fall into the category of (1)

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Mike Tintner
Ben, Well then so is S Kauffman's language unclear. I'll go with his definition in Chap 12 Reinventing the Sacred [all about algorithms and their impossibility for solving a whole string of human problems] What is an algorithm? The quick definition is an *effective procedure to calculate a

Re: [agi] Case-by-case Problem Solving (draft)

2008-09-18 Thread Mike Tintner
Matt, Thanks for reference. But it's still somewhat ambiguous. I could somewhat similarly outline a non-procedure procedure which might include steps like Think about the problem then Do something, anything - whatever first comes to mind and If that doesn't work, try something else. But as

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Ben Goertzel [EMAIL PROTECTED] wrote: Well, yes, and that difference is a distributed index, which has yet to be built. I extremely strongly disagree with the prior sentence ... I do not think that a distributed index is a sufficient architecture for powerful AGI at the

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, John LaMuth [EMAIL PROTECTED] wrote: You have completely left out the human element or friendly-type appeal How about a AGI personal assistant / tutor / PR interface Everyone should have one The market would be virtually unlimited ... That falls under the category

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Ben Goertzel
So perhaps you could name some applications of AGI that don't fall into the categories of (1) doing work or (2) augmenting your brain? 3) learning as much as possible 4) proving as many theorems as possible 5) figuring out how to improve human life as much as possible Of course, if you

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread Matt Mahoney
--- On Thu, 9/18/08, Trent Waddington [EMAIL PROTECTED] wrote: On Fri, Sep 19, 2008 at 11:34 AM, Matt Mahoney [EMAIL PROTECTED] wrote: So perhaps you could name some applications of AGI that don't fall into the categories of (1) doing work or (2) augmenting your brain? Perhaps you could

Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source)

2008-09-18 Thread John LaMuth
- Original Message - From: Matt Mahoney [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Thursday, September 18, 2008 7:45 PM Subject: Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source) --- On Thu, 9/18/08, John LaMuth [EMAIL PROTECTED] wrote: You have

Repair Theory (was Re: Two goals of AGI (was Re: [agi] Re: [OpenCog] Re: Proprietary_Open_Source))

2008-09-18 Thread Steve Richfield
Mike, On 9/18/08, Mike Tintner [EMAIL PROTECTED] wrote: Steve:View #2 (mine, stated from your approximate viewpoint) is that simple programs (like Dr. Eliza) have in the past and will in the future do things that people aren't good at. This includes tasks that encroach on intelligence, e.g.